diff --git druid-handler/src/java/org/apache/hadoop/hive/druid/serde/DruidGroupByQueryRecordReader.java druid-handler/src/java/org/apache/hadoop/hive/druid/serde/DruidGroupByQueryRecordReader.java index 9e8b439..f0bdb9e 100644 --- druid-handler/src/java/org/apache/hadoop/hive/druid/serde/DruidGroupByQueryRecordReader.java +++ druid-handler/src/java/org/apache/hadoop/hive/druid/serde/DruidGroupByQueryRecordReader.java @@ -99,7 +99,7 @@ public boolean nextKeyValue() { indexes[i]--; for (int j = i + 1; j < indexes.length; j++) { indexes[j] = current.getDimension( - query.getDimensions().get(j).getDimension()).size() - 1; + query.getDimensions().get(j).getOutputName()).size() - 1; } return true; } @@ -110,7 +110,7 @@ public boolean nextKeyValue() { indexes = new int[query.getDimensions().size()]; for (int i = 0; i < query.getDimensions().size(); i++) { DimensionSpec ds = query.getDimensions().get(i); - indexes[i] = current.getDimension(ds.getDimension()).size() - 1; + indexes[i] = current.getDimension(ds.getOutputName()).size() - 1; } return true; } @@ -131,7 +131,7 @@ public DruidWritable getCurrentValue() throws IOException, InterruptedException // 2) The dimension columns for (int i = 0; i < query.getDimensions().size(); i++) { DimensionSpec ds = query.getDimensions().get(i); - List dims = current.getDimension(ds.getDimension()); + List dims = current.getDimension(ds.getOutputName()); if (dims.size() == 0) { // NULL value for dimension value.getValue().put(ds.getOutputName(), null); @@ -170,7 +170,7 @@ public boolean next(NullWritable key, DruidWritable value) { // 2) The dimension columns for (int i = 0; i < query.getDimensions().size(); i++) { DimensionSpec ds = query.getDimensions().get(i); - List dims = current.getDimension(ds.getDimension()); + List dims = current.getDimension(ds.getOutputName()); if (dims.size() == 0) { // NULL value for dimension value.getValue().put(ds.getOutputName(), null);