diff --git a/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java b/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java index 1348d8b..933d6af 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/parse/spark/SplitOpTreeForDPP.java @@ -19,7 +19,7 @@ package org.apache.hadoop.hive.ql.parse.spark; import java.util.ArrayList; -import java.util.HashSet; +import java.util.LinkedHashSet; import java.util.LinkedList; import java.util.List; import java.util.Set; @@ -140,7 +140,7 @@ public Object process(Node nd, Stack stack, NodeProcessorCtx procCtx, } //Find all pruningSinkSet in new roots - Set> sinkSet = new HashSet<>(); + Set> sinkSet = new LinkedHashSet<>(); for (Operator root : newRoots) { SparkUtilities.collectOp(sinkSet, root, SparkPartitionPruningSinkOperator.class); } diff --git a/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out b/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out index 26680f8..d476172 100644 --- a/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out +++ b/ql/src/test/results/clientpositive/spark/spark_dynamic_partition_pruning.q.out @@ -3700,7 +3700,7 @@ STAGE PLANS: partition key expr: ds Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE target column name: ds - target work: Map 4 + target work: Map 1 Select Operator expressions: _col0 (type: string) outputColumnNames: _col0 @@ -3714,7 +3714,7 @@ STAGE PLANS: partition key expr: ds Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE target column name: ds - target work: Map 1 + target work: Map 4 Reducer 13 Reduce Operator Tree: Group By Operator @@ -3743,7 +3743,7 @@ STAGE PLANS: partition key expr: ds Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE target column name: ds - target work: Map 4 + target work: Map 1 Select Operator expressions: _col0 (type: string) outputColumnNames: _col0 @@ -3757,7 +3757,7 @@ STAGE PLANS: partition key expr: ds Statistics: Num rows: 2 Data size: 368 Basic stats: COMPLETE Column stats: NONE target column name: ds - target work: Map 1 + target work: Map 4 Stage: Stage-1 Spark