commit d2a68268409e40c849397dead20251bb462fb5b4 Author: Andrew Sherman Date: Wed Oct 25 13:50:15 2017 -0700 HIVE-17935: set the default for hive.optimize.sort.dynamic.partition to true diff --git common/src/java/org/apache/hadoop/hive/conf/HiveConf.java common/src/java/org/apache/hadoop/hive/conf/HiveConf.java index bd25bc7cadbed01d5918f1523b4fd055641ea98d..8f90b7a7debd065ba53df3ce40cf9e52aee8b1e9 100644 --- common/src/java/org/apache/hadoop/hive/conf/HiveConf.java +++ common/src/java/org/apache/hadoop/hive/conf/HiveConf.java @@ -1612,7 +1612,7 @@ private static void populateLlapDaemonVarsSet(Set llapDaemonVarsSetLocal "Currently it only works with Apache Tez. This should always be set to true. \n" + "Since it is a new feature, it has been made configurable."), - HIVEOPTSORTDYNAMICPARTITION("hive.optimize.sort.dynamic.partition", false, + HIVEOPTSORTDYNAMICPARTITION("hive.optimize.sort.dynamic.partition", true, "When enabled dynamic partitioning column will be globally sorted.\n" + "This way we can keep only one record writer open for each partition value\n" + "in the reducer thereby reducing the memory pressure on reducers."), diff --git itests/hive-unit/src/test/java/org/apache/hive/jdbc/AbstractJdbcTriggersTest.java itests/hive-unit/src/test/java/org/apache/hive/jdbc/AbstractJdbcTriggersTest.java index 235e6c36ed5328fcc2fc8d7108d3ea9d8dfc73a7..2b1bea947e3de38e5e958927bcf612bb576c0246 100644 --- itests/hive-unit/src/test/java/org/apache/hive/jdbc/AbstractJdbcTriggersTest.java +++ itests/hive-unit/src/test/java/org/apache/hive/jdbc/AbstractJdbcTriggersTest.java @@ -72,6 +72,7 @@ public static void beforeTest() throws Exception { conf.setBoolVar(ConfVars.TEZ_EXEC_SUMMARY, true); conf.setBoolVar(ConfVars.HIVE_STRICT_CHECKS_CARTESIAN, false); conf.setVar(ConfVars.LLAP_IO_MEMORY_MODE, "none"); + conf.setBoolVar(ConfVars.HIVEOPTSORTDYNAMICPARTITION, false); conf.addResource(new URL("file://" + new File(confDir).toURI().getPath() + "/tez-site.xml")); @@ -141,7 +142,7 @@ void runQueryWithTrigger(final String query, final List setCmds, if (expect == null) { assertNull("Expected query to succeed", throwable[0]); } else { - assertNotNull("Expected non-null throwable", throwable[0]); + assertNotNull("Expected non-null throwable for " + expect, throwable[0]); assertEquals(SQLException.class, throwable[0].getClass()); assertTrue(expect + " is not contained in " + throwable[0].getMessage(), throwable[0].getMessage().contains(expect)); diff --git ql/src/test/results/clientnegative/ct_noperm_loc.q.out ql/src/test/results/clientnegative/ct_noperm_loc.q.out index 0b8182a69b1411b49c0f20f513af906ba7b8f510..e40dae32ea50bc1f4b89552300cd8c1e87e5fde2 100644 --- ql/src/test/results/clientnegative/ct_noperm_loc.q.out +++ ql/src/test/results/clientnegative/ct_noperm_loc.q.out @@ -1 +1,11 @@ #### A masked pattern was here #### +PREHOOK: type: CREATETABLE +#### A masked pattern was here #### +PREHOOK: Output: database:default +PREHOOK: Output: default@foo0 +#### A masked pattern was here #### +POSTHOOK: type: CREATETABLE +#### A masked pattern was here #### +POSTHOOK: Output: database:default +POSTHOOK: Output: default@foo0 +FAILED: SemanticException [Error 10043]: Either list of columns or a custom serializer should be specified diff --git ql/src/test/results/clientpositive/annotate_stats_part.q.out ql/src/test/results/clientpositive/annotate_stats_part.q.out index 3a94a6a4e33a715fd54e4a31b38bfb035da4dcc7..643e724674804a15a5aa37a1019ef5b0a981a45e 100644 --- ql/src/test/results/clientpositive/annotate_stats_part.q.out +++ ql/src/test/results/clientpositive/annotate_stats_part.q.out @@ -90,7 +90,7 @@ STAGE PLANS: Processor Tree: TableScan alias: loc_orc - Statistics: Num rows: 6 Data size: 3060 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 6 Data size: 3059 Basic stats: COMPLETE Column stats: PARTIAL Select Operator expressions: state (type: string), locid (type: int), zip (type: bigint), year (type: string) outputColumnNames: _col0, _col1, _col2, _col3 diff --git ql/src/test/results/clientpositive/autoColumnStats_1.q.out ql/src/test/results/clientpositive/autoColumnStats_1.q.out index e04eaf45308d3a80af2944592fed3f12eadf1b21..89f787737131e6ecb75dbd9f0899d38416dd009a 100644 --- ql/src/test/results/clientpositive/autoColumnStats_1.q.out +++ ql/src/test/results/clientpositive/autoColumnStats_1.q.out @@ -489,11 +489,11 @@ STAGE PLANS: Processor Tree: TableScan alias: nzhang_part14 - Statistics: Num rows: 6 Data size: 516 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 6 Data size: 1116 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 6 Data size: 516 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 6 Data size: 1116 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: drop table src5 @@ -545,11 +545,11 @@ STAGE PLANS: Processor Tree: TableScan alias: nzhang_part14 - Statistics: Num rows: 11 Data size: 946 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 11 Data size: 2049 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 11 Data size: 946 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 11 Data size: 2049 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: create table alter5 ( col1 string ) partitioned by (dt string) @@ -1340,11 +1340,11 @@ STAGE PLANS: Processor Tree: TableScan alias: a - Statistics: Num rows: 2000 Data size: 174000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 371032 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 2000 Data size: 174000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 371032 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: explain select value from b @@ -1361,11 +1361,11 @@ STAGE PLANS: Processor Tree: TableScan alias: b - Statistics: Num rows: 1000 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1000 Data size: 185608 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: value (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1000 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1000 Data size: 185608 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: explain select key from b @@ -1382,11 +1382,11 @@ STAGE PLANS: Processor Tree: TableScan alias: b - Statistics: Num rows: 1000 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1000 Data size: 185608 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1000 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1000 Data size: 185608 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: explain select value from c @@ -1403,11 +1403,11 @@ STAGE PLANS: Processor Tree: TableScan alias: c - Statistics: Num rows: 2000 Data size: 182000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 371032 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: value (type: string) outputColumnNames: _col0 - Statistics: Num rows: 2000 Data size: 182000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 371032 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: explain select key from c @@ -1424,10 +1424,10 @@ STAGE PLANS: Processor Tree: TableScan alias: c - Statistics: Num rows: 2000 Data size: 174000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 371032 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 2000 Data size: 174000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 371032 Basic stats: COMPLETE Column stats: NONE ListSink diff --git ql/src/test/results/clientpositive/autoColumnStats_2.q.out ql/src/test/results/clientpositive/autoColumnStats_2.q.out index b209ff0dccd76b86f3c0ef7359f7d5d750b713e8..52ce6f76b0ebce2645cff89e632c8389ffcd7b81 100644 --- ql/src/test/results/clientpositive/autoColumnStats_2.q.out +++ ql/src/test/results/clientpositive/autoColumnStats_2.q.out @@ -487,11 +487,11 @@ STAGE PLANS: Processor Tree: TableScan alias: nzhang_part14 - Statistics: Num rows: 6 Data size: 516 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 6 Data size: 1116 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 6 Data size: 516 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 6 Data size: 1116 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: drop table src5 @@ -543,11 +543,11 @@ STAGE PLANS: Processor Tree: TableScan alias: nzhang_part14 - Statistics: Num rows: 11 Data size: 946 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 11 Data size: 2049 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 11 Data size: 946 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 11 Data size: 2049 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: drop table alter5 @@ -1464,11 +1464,11 @@ STAGE PLANS: Processor Tree: TableScan alias: a - Statistics: Num rows: 2000 Data size: 174000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 371032 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 2000 Data size: 174000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 371032 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: explain select value from b @@ -1485,11 +1485,11 @@ STAGE PLANS: Processor Tree: TableScan alias: b - Statistics: Num rows: 1000 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1000 Data size: 185608 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: value (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1000 Data size: 91000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1000 Data size: 185608 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: explain select key from b @@ -1506,11 +1506,11 @@ STAGE PLANS: Processor Tree: TableScan alias: b - Statistics: Num rows: 1000 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1000 Data size: 185608 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1000 Data size: 87000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 1000 Data size: 185608 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: explain select value from c @@ -1527,11 +1527,11 @@ STAGE PLANS: Processor Tree: TableScan alias: c - Statistics: Num rows: 2000 Data size: 182000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 371032 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: value (type: string) outputColumnNames: _col0 - Statistics: Num rows: 2000 Data size: 182000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 371032 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: explain select key from c @@ -1548,10 +1548,10 @@ STAGE PLANS: Processor Tree: TableScan alias: c - Statistics: Num rows: 2000 Data size: 174000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 371032 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: string) outputColumnNames: _col0 - Statistics: Num rows: 2000 Data size: 174000 Basic stats: COMPLETE Column stats: COMPLETE + Statistics: Num rows: 2000 Data size: 371032 Basic stats: COMPLETE Column stats: NONE ListSink diff --git ql/src/test/results/clientpositive/autoColumnStats_6.q.out ql/src/test/results/clientpositive/autoColumnStats_6.q.out index 70788fde9f356bc55a1c95a52aa85c52d367b214..6e3c62be5b6f9ba01b4f87766176c4218a91c9f3 100644 --- ql/src/test/results/clientpositive/autoColumnStats_6.q.out +++ ql/src/test/results/clientpositive/autoColumnStats_6.q.out @@ -26,13 +26,8 @@ POSTHOOK: query: EXPLAIN INSERT OVERWRITE TABLE orcfile_merge2a PARTITION (one=' POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-0 depends on stages: Stage-1 Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 STAGE PLANS: Stage: Stage-1 @@ -45,57 +40,26 @@ STAGE PLANS: expressions: UDFToInteger(key) (type: int), value (type: string), (hash(key) pmod 10) (type: int), (hash(value) pmod 10) (type: int) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: int), _col3 (type: int) + sort order: ++ + Map-reduce partition columns: _col2 (type: int), _col3 (type: int) Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat - serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde - name: default.orcfile_merge2a - Select Operator - expressions: _col0 (type: int), _col1 (type: string), '1' (type: string), UDFToString(_col2) (type: string), UDFToString(_col3) (type: string) - outputColumnNames: key, value, one, two, three - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: one (type: string), two (type: string), three (type: string) - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - sort order: +++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string), _col2 (type: string) - Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - value expressions: _col3 (type: struct), _col4 (type: struct) + value expressions: _col0 (type: int), _col1 (type: string) Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string), KEY._col2 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col3 (type: struct), _col4 (type: struct), _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3, _col4 - Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: string), KEY._col2 (type: int), KEY._col3 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat + output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat + serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde + name: default.orcfile_merge2a Stage: Stage-0 Move Operator @@ -119,26 +83,6 @@ STAGE PLANS: Column Types: int, string Table: default.orcfile_merge2a - Stage: Stage-3 - Merge File Operator - Map Operator Tree: - ORC File Merge Operator - merge level: stripe - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - - Stage: Stage-5 - Merge File Operator - Map Operator Tree: - ORC File Merge Operator - merge level: stripe - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: INSERT OVERWRITE TABLE orcfile_merge2a PARTITION (one='1', two, three) SELECT key, value, PMOD(HASH(key), 10) as two, PMOD(HASH(value), 10) as three diff --git ql/src/test/results/clientpositive/autoColumnStats_8.q.out ql/src/test/results/clientpositive/autoColumnStats_8.q.out index 1d7e966a317e7134ae377e5653f73862d97eba4f..3f77cfc32ca54e5692f1af06b77b9f1d5354b32b 100644 --- ql/src/test/results/clientpositive/autoColumnStats_8.q.out +++ ql/src/test/results/clientpositive/autoColumnStats_8.q.out @@ -47,9 +47,9 @@ STAGE DEPENDENCIES: Stage-2 is a root stage Stage-0 depends on stages: Stage-2 Stage-3 depends on stages: Stage-0 - Stage-1 depends on stages: Stage-2 - Stage-4 depends on stages: Stage-1, Stage-5 - Stage-5 depends on stages: Stage-2 + Stage-4 depends on stages: Stage-2 + Stage-1 depends on stages: Stage-4 + Stage-5 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-2 @@ -67,54 +67,15 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.nzhang_part8 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct nzhang_part8 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part8 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string), _col3 (type: string) - outputColumnNames: key, value, ds, hr + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + null sort order: aa + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: ds (type: string), hr (type: string) - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - Reduce Output Operator - key expressions: _col0 (type: string), _col1 (type: string) - null sort order: aa - sort order: ++ - Map-reduce partition columns: _col0 (type: string), _col1 (type: string) - Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) - auto parallelism: false + tag: -1 + value expressions: _col0 (type: string), _col1 (type: string) + auto parallelism: false Filter Operator isSamplingPred: false predicate: (ds > '2008-04-08') (type: boolean) @@ -125,62 +86,22 @@ STAGE PLANS: Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - GlobalTableId: 2 + GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-12-31/ - Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: - bucket_count -1 column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.nzhang_part8 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct nzhang_part8 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part8 + columns _col0,_col1,_col2 + columns.types string,string,string + escape.delim \ + serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 - GatherStats: true + GatherStats: false MultiFileSpray: false - Select Operator - expressions: _col0 (type: string), _col1 (type: string), _col2 (type: string) - outputColumnNames: key, value, hr - Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - Group By Operator - aggregations: compute_stats(key, 'hll'), compute_stats(value, 'hll') - keys: '2008-12-31' (type: string), hr (type: string) - mode: hash - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - GlobalTableId: 0 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types string,string,struct,struct - escape.delim \ - serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false Path -> Alias: #### A masked pattern was here #### Path -> Partition: @@ -383,38 +304,40 @@ STAGE PLANS: /srcpart/ds=2008-04-09/hr=12 [srcpart] Needs Tagging: false Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), _col0 (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - GlobalTableId: 0 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + GlobalTableId: 1 #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE + Dp Sort State: PARTITION_SORTED + NumFilesPerFileSink: 1 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.nzhang_part8 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct nzhang_part8 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part8 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Stage: Stage-0 Move Operator @@ -449,57 +372,19 @@ STAGE PLANS: Basic Stats Work: #### A masked pattern was here #### - Stage: Stage-1 - Move Operator - tables: - partition: - ds 2008-12-31 - hr - replace: true -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.nzhang_part8 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct nzhang_part8 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part8 - Stage: Stage-4 - Stats Work - Basic Stats Work: -#### A masked pattern was here #### - Column Stats Desc: - Columns: key, value - Column Types: string, string - Table: default.nzhang_part8 - Is Table Level Stats: false - - Stage: Stage-5 Map Reduce Map Operator Tree: TableScan GatherStats: false Reduce Output Operator - key expressions: '2008-12-31' (type: string), _col1 (type: string) - null sort order: aa - sort order: ++ - Map-reduce partition columns: '2008-12-31' (type: string), _col1 (type: string) + key expressions: _col2 (type: string) + null sort order: a + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE tag: -1 - value expressions: _col2 (type: struct), _col3 (type: struct) + value expressions: _col0 (type: string), _col1 (type: string) auto parallelism: false Path -> Alias: #### A masked pattern was here #### @@ -511,8 +396,8 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types string,string,struct,struct + columns _col0,_col1,_col2 + columns.types string,string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe @@ -521,8 +406,8 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: column.name.delimiter , - columns _col0,_col1,_col2,_col3 - columns.types string,string,struct,struct + columns _col0,_col1,_col2 + columns.types string,string,string escape.delim \ serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe @@ -530,38 +415,79 @@ STAGE PLANS: #### A masked pattern was here #### Needs Tagging: false Reduce Operator Tree: - Group By Operator - aggregations: compute_stats(VALUE._col0), compute_stats(VALUE._col1) - keys: '2008-12-31' (type: string), KEY._col1 (type: string) - mode: mergepartial - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col2 (type: struct), _col3 (type: struct), '2008-12-31' (type: string), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - GlobalTableId: 0 + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + GlobalTableId: 2 +#### A masked pattern was here #### + Dp Sort State: PARTITION_SORTED + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-12-31/ + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.nzhang_part8 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct nzhang_part8 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part8 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + + Stage: Stage-1 + Move Operator + tables: + partition: + ds 2008-12-31 + hr + replace: true +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - NumFilesPerFileSink: 1 - Statistics: Num rows: 333 Data size: 3537 Basic stats: COMPLETE Column stats: NONE + name default.nzhang_part8 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct nzhang_part8 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.SequenceFileInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat - properties: - columns _col0,_col1,_col2,_col3 - columns.types struct:struct:string:string - escape.delim \ - hive.serialization.extend.additional.nesting.levels true - serialization.escape.crlf true - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - TotalFiles: 1 - GatherStats: false - MultiFileSpray: false + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part8 + + Stage: Stage-5 + Stats Work + Basic Stats Work: +#### A masked pattern was here #### + Column Stats Desc: + Columns: key, value + Column Types: string, string + Table: default.nzhang_part8 + Is Table Level Stats: false PREHOOK: query: from srcpart insert overwrite table nzhang_part8 partition (ds, hr) select key, value, ds, hr where ds <= '2008-04-08' diff --git ql/src/test/results/clientpositive/delete_all_partitioned.q.out ql/src/test/results/clientpositive/delete_all_partitioned.q.out index 90f8753687b23fc1910852995747a578cb3cb47b..cd9b14340d6cedf8f175cb6532719f97df538fe6 100644 --- ql/src/test/results/clientpositive/delete_all_partitioned.q.out +++ ql/src/test/results/clientpositive/delete_all_partitioned.q.out @@ -84,3 +84,16 @@ POSTHOOK: Input: default@acid_dap POSTHOOK: Input: default@acid_dap@ds=today POSTHOOK: Input: default@acid_dap@ds=tomorrow #### A masked pattern was here #### +-1072076362 2uLyD28144vklju213J1mr today +-1069736047 k17Am8uPHWk02cEf1jet today +-1070551679 iUR3Q today +-1070883071 0ruyd6Y50JpdGRf6HqD today +-1071363017 Anj0oF today +-1072081801 dPkN74F7 today +-1072910839 0iqrc5 today +-1073279343 oj1YrV5Wa today +6981 YdG61y00526u5 tomorrow +6981 o5mb0QP5Y48Qd4vdB0 tomorrow +6981 o4lvY20511w0EOX3P3I82p63 tomorrow +6981 NULL tomorrow +6981 K630vaVf tomorrow diff --git ql/src/test/results/clientpositive/extrapolate_part_stats_partial.q.out ql/src/test/results/clientpositive/extrapolate_part_stats_partial.q.out index 5f74166e894be125729fcacd06045e24fef64c87..31f96ff3a27bee68ea27371a2550c31cd3774786 100644 --- ql/src/test/results/clientpositive/extrapolate_part_stats_partial.q.out +++ ql/src/test/results/clientpositive/extrapolate_part_stats_partial.q.out @@ -251,7 +251,7 @@ STAGE PLANS: serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 410 + totalSize 409 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -296,7 +296,7 @@ STAGE PLANS: serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 419 + totalSize 420 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -454,7 +454,7 @@ STAGE PLANS: serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 410 + totalSize 409 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -499,7 +499,7 @@ STAGE PLANS: serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 419 + totalSize 420 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -685,7 +685,7 @@ STAGE PLANS: serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 410 + totalSize 409 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -730,7 +730,7 @@ STAGE PLANS: serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 419 + totalSize 420 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -888,7 +888,7 @@ STAGE PLANS: serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 410 + totalSize 409 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -933,7 +933,7 @@ STAGE PLANS: serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 419 + totalSize 420 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde diff --git ql/src/test/results/clientpositive/implicit_cast_during_insert.q.out ql/src/test/results/clientpositive/implicit_cast_during_insert.q.out index 1fa9e94aa7ef7d23949bf2d48ecd6b6a81dee14a..668400faa09d415277e258238035c12ea619a510 100644 --- ql/src/test/results/clientpositive/implicit_cast_during_insert.q.out +++ ql/src/test/results/clientpositive/implicit_cast_during_insert.q.out @@ -39,17 +39,18 @@ STAGE PLANS: Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col2 (type: string) - sort order: + + sort order: ++ Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE - value expressions: _col1 (type: string) + value expressions: UDFToInteger(_col2) (type: int), _col1 (type: string) Reduce Operator Tree: Select Operator - expressions: UDFToInteger(KEY.reducesinkkey0) (type: int), VALUE._col0 (type: string), KEY.reducesinkkey0 (type: string) + expressions: VALUE._col0 (type: int), VALUE._col1 (type: string), KEY._col2 (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false + Dp Sort State: PARTITION_SORTED Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat diff --git ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out index cba50a758d81d93dffd600d9cec5e17c18a76d8d..e8a9866fb77c58452263e406a6aae953ffc9e5e1 100644 --- ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out +++ ql/src/test/results/clientpositive/infer_bucket_sort_dyn_part.q.out @@ -197,7 +197,7 @@ Partition Parameters: numFiles 1 numRows 309 rawDataSize 1173 - totalSize 1342 + totalSize 1381 #### A masked pattern was here #### # Storage Information @@ -260,7 +260,7 @@ Partition Parameters: numFiles 1 numRows 155 rawDataSize 586 - totalSize 719 + totalSize 722 #### A masked pattern was here #### # Storage Information @@ -268,9 +268,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: DESCRIBE FORMATTED test_table PARTITION (ds='2008-04-08', hr='12') @@ -298,7 +298,7 @@ Partition Parameters: numFiles 1 numRows 154 rawDataSize 591 - totalSize 722 + totalSize 741 #### A masked pattern was here #### # Storage Information @@ -306,9 +306,9 @@ SerDe Library: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe InputFormat: org.apache.hadoop.hive.ql.io.RCFileInputFormat OutputFormat: org.apache.hadoop.hive.ql.io.RCFileOutputFormat Compressed: No -Num Buckets: 1 -Bucket Columns: [key] -Sort Columns: [Order(col:key, order:1)] +Num Buckets: -1 +Bucket Columns: [] +Sort Columns: [] Storage Desc Params: serialization.format 1 PREHOOK: query: CREATE TABLE srcpart_merge_dp LIKE srcpart @@ -412,13 +412,14 @@ GROUP BY key) a POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 + Stage-2 depends on stages: Stage-1 + Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 + Stage-3 depends on stages: Stage-0 + Stage-4 + Stage-6 + Stage-7 depends on stages: Stage-6 STAGE PLANS: Stage: Stage-1 @@ -456,17 +457,40 @@ STAGE PLANS: Statistics: Num rows: 58 Data size: 5812 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 58 Data size: 5812 Basic stats: COMPLETE Column stats: NONE table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.test_table + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Stage: Stage-7 + Stage: Stage-2 + Map Reduce + Map Operator Tree: + TableScan + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 58 Data size: 5812 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 58 Data size: 5812 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 58 Data size: 5812 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.test_table + + Stage: Stage-8 Conditional Operator - Stage: Stage-4 + Stage: Stage-5 Move Operator files: hdfs directory: true @@ -485,25 +509,25 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.test_table - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: - Stage: Stage-3 + Stage: Stage-4 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-6 + Stage: Stage-7 Move Operator files: hdfs directory: true @@ -559,7 +583,7 @@ Partition Parameters: numFiles 1 numRows 4 rawDataSize 14 - totalSize 115 + totalSize 94 #### A masked pattern was here #### # Storage Information @@ -594,10 +618,10 @@ Table: test_table #### A masked pattern was here #### Partition Parameters: COLUMN_STATS_ACCURATE {\"BASIC_STATS\":\"true\"} - numFiles 2 + numFiles 1 numRows 305 rawDataSize 1163 - totalSize 1427 + totalSize 1346 #### A masked pattern was here #### # Storage Information diff --git ql/src/test/results/clientpositive/insert_into6.q.out ql/src/test/results/clientpositive/insert_into6.q.out index 964e259771d619197879410aafde42c0fea242f1..e369831a25440b6a0cc335112452880c7929fb81 100644 --- ql/src/test/results/clientpositive/insert_into6.q.out +++ ql/src/test/results/clientpositive/insert_into6.q.out @@ -134,13 +134,8 @@ POSTHOOK: query: EXPLAIN INSERT INTO TABLE insert_into6b PARTITION (ds) POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-0 depends on stages: Stage-1 Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 STAGE PLANS: Stage: Stage-1 @@ -153,23 +148,26 @@ STAGE PLANS: expressions: key (type: int), value (type: string), ds (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 250 Data size: 2680 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 250 Data size: 2680 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.insert_into6b - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + value expressions: _col0 (type: int), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 250 Data size: 2680 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 250 Data size: 2680 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.insert_into6b Stage: Stage-0 Move Operator @@ -187,36 +185,6 @@ STAGE PLANS: Stats Work Basic Stats Work: - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.insert_into6b - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.insert_into6b - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: INSERT INTO TABLE insert_into6b PARTITION (ds) SELECT * FROM insert_into6a PREHOOK: type: QUERY PREHOOK: Input: default@insert_into6a diff --git ql/src/test/results/clientpositive/llap/delete_all_partitioned.q.out ql/src/test/results/clientpositive/llap/delete_all_partitioned.q.out index 90f8753687b23fc1910852995747a578cb3cb47b..32c2a0e3baa10d136c4281da9bdc636c4ec4b9af 100644 --- ql/src/test/results/clientpositive/llap/delete_all_partitioned.q.out +++ ql/src/test/results/clientpositive/llap/delete_all_partitioned.q.out @@ -84,3 +84,16 @@ POSTHOOK: Input: default@acid_dap POSTHOOK: Input: default@acid_dap@ds=today POSTHOOK: Input: default@acid_dap@ds=tomorrow #### A masked pattern was here #### +-1072076362 2uLyD28144vklju213J1mr today +-1069736047 k17Am8uPHWk02cEf1jet today +-1070551679 iUR3Q today +-1070883071 0ruyd6Y50JpdGRf6HqD today +-1071363017 Anj0oF today +-1072081801 dPkN74F7 today +-1072910839 0iqrc5 today +-1073279343 oj1YrV5Wa today +6981 a3EhVU6Wuy7ycJ7wY7h2gv tomorrow +6981 1FNNhmiFLGw425NA13g tomorrow +6981 sF2CRfgt2K tomorrow +6981 4KhrrQ0nJ7bMNTvhSCA tomorrow +6981 Y5x3JuI3M8jngv5N tomorrow diff --git ql/src/test/results/clientpositive/llap/dp_counter_mm.q.out ql/src/test/results/clientpositive/llap/dp_counter_mm.q.out index 662c02a778df37775f3cbf1153596ba871a632a2..72c6387ce0db399b0fdbe5b9a9b3a17ceb4a1bb1 100644 --- ql/src/test/results/clientpositive/llap/dp_counter_mm.q.out +++ ql/src/test/results/clientpositive/llap/dp_counter_mm.q.out @@ -21,6 +21,7 @@ Stage-1 HIVE COUNTERS: DESERIALIZE_ERRORS: 0 RECORDS_IN_Map_1: 500 RECORDS_OUT_1_default.src2: 84 + RECORDS_OUT_INTERMEDIATE_Map_1: 84 PREHOOK: query: insert into table src2 partition (value) select * from src where key < 200 PREHOOK: type: QUERY PREHOOK: Input: default@src @@ -32,6 +33,7 @@ Stage-1 HIVE COUNTERS: DESERIALIZE_ERRORS: 0 RECORDS_IN_Map_1: 500 RECORDS_OUT_1_default.src2: 189 + RECORDS_OUT_INTERMEDIATE_Map_1: 189 PREHOOK: query: drop table src2 PREHOOK: type: DROPTABLE PREHOOK: Input: default@src2 @@ -51,6 +53,7 @@ Stage-1 HIVE COUNTERS: DESERIALIZE_ERRORS: 0 RECORDS_IN_Map_1: 500 RECORDS_OUT_1_default.src2: 189 + RECORDS_OUT_INTERMEDIATE_Map_1: 189 PREHOOK: query: insert into table src2 partition (value) select * from src where key < 300 PREHOOK: type: QUERY PREHOOK: Input: default@src @@ -62,6 +65,7 @@ Stage-1 HIVE COUNTERS: DESERIALIZE_ERRORS: 0 RECORDS_IN_Map_1: 500 RECORDS_OUT_1_default.src2: 292 + RECORDS_OUT_INTERMEDIATE_Map_1: 292 PREHOOK: query: drop table src2 PREHOOK: type: DROPTABLE PREHOOK: Input: default@src2 @@ -91,6 +95,7 @@ Stage-2 HIVE COUNTERS: RECORDS_IN_Map_1: 500 RECORDS_OUT_1_default.src2: 84 RECORDS_OUT_2_default.src3: 105 + RECORDS_OUT_INTERMEDIATE_Map_1: 189 PREHOOK: query: from src insert into table src2 partition (value) select * where key < 100 insert into table src3 partition (value) select * where key >= 100 and key < 300 @@ -106,6 +111,7 @@ Stage-2 HIVE COUNTERS: RECORDS_IN_Map_1: 500 RECORDS_OUT_1_default.src2: 84 RECORDS_OUT_2_default.src3: 208 + RECORDS_OUT_INTERMEDIATE_Map_1: 292 PREHOOK: query: drop table src2 PREHOOK: type: DROPTABLE PREHOOK: Input: default@src2 @@ -128,8 +134,10 @@ Stage-1 HIVE COUNTERS: CREATED_FILES: 121 DESERIALIZE_ERRORS: 0 RECORDS_IN_Map_1: 500 - RECORDS_IN_Map_3: 500 + RECORDS_IN_Map_4: 500 RECORDS_OUT_1_default.src2: 189 + RECORDS_OUT_INTERMEDIATE_Map_1: 84 + RECORDS_OUT_INTERMEDIATE_Map_4: 105 PREHOOK: query: insert into table src2 partition (value) select temps.* from ( select * from src where key < 100 @@ -144,5 +152,7 @@ Stage-1 HIVE COUNTERS: CREATED_FILES: 184 DESERIALIZE_ERRORS: 0 RECORDS_IN_Map_1: 500 - RECORDS_IN_Map_3: 500 + RECORDS_IN_Map_4: 500 RECORDS_OUT_1_default.src2: 292 + RECORDS_OUT_INTERMEDIATE_Map_1: 84 + RECORDS_OUT_INTERMEDIATE_Map_4: 208 diff --git ql/src/test/results/clientpositive/llap/dp_counter_non_mm.q.out ql/src/test/results/clientpositive/llap/dp_counter_non_mm.q.out index 85f9b9d7a3779c307c0f2c0a8b2780b64f9a535d..0b799043e0bfcc409205c724b7d11a3bf751805d 100644 --- ql/src/test/results/clientpositive/llap/dp_counter_non_mm.q.out +++ ql/src/test/results/clientpositive/llap/dp_counter_non_mm.q.out @@ -21,6 +21,7 @@ Stage-1 HIVE COUNTERS: DESERIALIZE_ERRORS: 0 RECORDS_IN_Map_1: 500 RECORDS_OUT_1_default.src2: 84 + RECORDS_OUT_INTERMEDIATE_Map_1: 84 PREHOOK: query: insert into table src2 partition (value) select * from src where key < 200 PREHOOK: type: QUERY PREHOOK: Input: default@src @@ -32,6 +33,7 @@ Stage-1 HIVE COUNTERS: DESERIALIZE_ERRORS: 0 RECORDS_IN_Map_1: 500 RECORDS_OUT_1_default.src2: 189 + RECORDS_OUT_INTERMEDIATE_Map_1: 189 PREHOOK: query: drop table src2 PREHOOK: type: DROPTABLE PREHOOK: Input: default@src2 @@ -51,6 +53,7 @@ Stage-1 HIVE COUNTERS: DESERIALIZE_ERRORS: 0 RECORDS_IN_Map_1: 500 RECORDS_OUT_1_default.src2: 189 + RECORDS_OUT_INTERMEDIATE_Map_1: 189 PREHOOK: query: insert into table src2 partition (value) select * from src where key < 300 PREHOOK: type: QUERY PREHOOK: Input: default@src @@ -62,6 +65,7 @@ Stage-1 HIVE COUNTERS: DESERIALIZE_ERRORS: 0 RECORDS_IN_Map_1: 500 RECORDS_OUT_1_default.src2: 292 + RECORDS_OUT_INTERMEDIATE_Map_1: 292 PREHOOK: query: drop table src2 PREHOOK: type: DROPTABLE PREHOOK: Input: default@src2 @@ -91,6 +95,7 @@ Stage-2 HIVE COUNTERS: RECORDS_IN_Map_1: 500 RECORDS_OUT_1_default.src2: 84 RECORDS_OUT_2_default.src3: 105 + RECORDS_OUT_INTERMEDIATE_Map_1: 189 PREHOOK: query: from src insert into table src2 partition (value) select * where key < 100 insert into table src3 partition (value) select * where key >= 100 and key < 300 @@ -106,6 +111,7 @@ Stage-2 HIVE COUNTERS: RECORDS_IN_Map_1: 500 RECORDS_OUT_1_default.src2: 84 RECORDS_OUT_2_default.src3: 208 + RECORDS_OUT_INTERMEDIATE_Map_1: 292 PREHOOK: query: drop table src2 PREHOOK: type: DROPTABLE PREHOOK: Input: default@src2 @@ -128,8 +134,10 @@ Stage-1 HIVE COUNTERS: CREATED_FILES: 121 DESERIALIZE_ERRORS: 0 RECORDS_IN_Map_1: 500 - RECORDS_IN_Map_3: 500 + RECORDS_IN_Map_4: 500 RECORDS_OUT_1_default.src2: 189 + RECORDS_OUT_INTERMEDIATE_Map_1: 84 + RECORDS_OUT_INTERMEDIATE_Map_4: 105 PREHOOK: query: insert into table src2 partition (value) select temps.* from ( select * from src where key < 100 @@ -144,5 +152,7 @@ Stage-1 HIVE COUNTERS: CREATED_FILES: 184 DESERIALIZE_ERRORS: 0 RECORDS_IN_Map_1: 500 - RECORDS_IN_Map_3: 500 + RECORDS_IN_Map_4: 500 RECORDS_OUT_1_default.src2: 292 + RECORDS_OUT_INTERMEDIATE_Map_1: 84 + RECORDS_OUT_INTERMEDIATE_Map_4: 208 diff --git ql/src/test/results/clientpositive/llap/extrapolate_part_stats_partial_ndv.q.out ql/src/test/results/clientpositive/llap/extrapolate_part_stats_partial_ndv.q.out index 1ac5977898840e313e8408ee76b3aa473e5138a1..b8ca3ec56d7a4a74834d39a4e5d83d50f37f39a8 100644 --- ql/src/test/results/clientpositive/llap/extrapolate_part_stats_partial_ndv.q.out +++ ql/src/test/results/clientpositive/llap/extrapolate_part_stats_partial_ndv.q.out @@ -375,7 +375,7 @@ STAGE PLANS: serialization.ddl struct loc_orc_1d { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 580 + totalSize 582 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde @@ -748,7 +748,7 @@ STAGE PLANS: serialization.ddl struct loc_orc_1d { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 580 + totalSize 582 #### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde diff --git ql/src/test/results/clientpositive/llap/lineage3.q.out ql/src/test/results/clientpositive/llap/lineage3.q.out index 66cc6ad5a0b14e5652482fa6faef771446f0cfc9..5c01f45a229068abfc421e7fbdea960b58ff577f 100644 --- ql/src/test/results/clientpositive/llap/lineage3.q.out +++ ql/src/test/results/clientpositive/llap/lineage3.q.out @@ -370,7 +370,7 @@ PREHOOK: query: create table src_dp1 (f string, w string, m int) PREHOOK: type: CREATETABLE PREHOOK: Output: database:default PREHOOK: Output: default@src_dp1 -Warning: Shuffle Join MERGEJOIN[19][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product +Warning: Shuffle Join MERGEJOIN[27][tables = [$hdt$_0, $hdt$_1]] in Stage 'Reducer 2' is a cross product PREHOOK: query: from src_dp, src_dp1 insert into dest_dp1 partition (year) select first, word, year insert into dest_dp2 partition (y, m) select first, word, year, month diff --git ql/src/test/results/clientpositive/llap/llap_acid.q.out ql/src/test/results/clientpositive/llap/llap_acid.q.out index 38889b93ba840657961a81e73a94a8f826369c64..d326585f3deeebff2fb63800ba898e093d59f64a 100644 --- ql/src/test/results/clientpositive/llap/llap_acid.q.out +++ ql/src/test/results/clientpositive/llap/llap_acid.q.out @@ -248,12 +248,14 @@ STAGE PLANS: native: true projectedOutputColumnNums: [5, 2, 3, 4] Reduce Sink Vectorization: + bucketColumnNums: [5] className: VectorReduceSinkObjectHashOperator - keyColumnNums: [5] + keyColumnNums: [4, 6, 5] + keyExpressions: ConstantVectorExpression(val _bucket_number) -> 6:string native: true nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true - partitionColumnNums: [6] - valueColumnNums: [2, 3, 4] + partitionColumnNums: [4] + valueColumnNums: [2, 3] Execution mode: vectorized, llap LLAP IO: may be used (ACID table) Map Vectorization: @@ -272,28 +274,28 @@ STAGE PLANS: neededVirtualColumns: [ROWID] partitionColumnCount: 1 partitionColumns: csmallint:smallint - scratchColumnTypeNames: [bigint] + scratchColumnTypeNames: [string] Reducer 2 Execution mode: vectorized, llap Reduce Vectorization: enabled: true enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true - reduceColumnNullOrder: a - reduceColumnSortOrder: + + reduceColumnNullOrder: aaa + reduceColumnSortOrder: +++ allNative: false usesVectorUDFAdaptor: false vectorized: true rowBatchContext: - dataColumnCount: 4 - dataColumns: KEY.reducesinkkey0:struct, VALUE._col1:float, VALUE._col2:double, VALUE._col3:smallint + dataColumnCount: 5 + dataColumns: KEY._col5:smallint, KEY.'_bucket_number':string, KEY._col0:struct, VALUE._col3:float, VALUE._col4:double partitionColumnCount: 0 scratchColumnTypeNames: [bigint, bigint] Reduce Operator Tree: Select Vectorization: className: VectorSelectOperator native: true - projectedOutputColumnNums: [0, 4, 5, 1, 2, 3] - selectExpressions: ConstantVectorExpression(val 1) -> 4:int, ConstantVectorExpression(val 2) -> 5:bigint + projectedOutputColumnNums: [2, 5, 6, 3, 4, 0, 1] + selectExpressions: ConstantVectorExpression(val 1) -> 5:int, ConstantVectorExpression(val 2) -> 6:bigint File Sink Vectorization: className: VectorFileSinkOperator native: false @@ -426,7 +428,7 @@ POSTHOOK: Input: default@orc_llap@csmallint=3 #### A masked pattern was here #### -285355633 1 -1241163445 -109813638 1 -58941842 -1 1 2 +1 1 1 164554497 1 1161977292 199879534 1 123351087 246423894 1 -1645852809 diff --git ql/src/test/results/clientpositive/llap/llap_acid_fast.q.out ql/src/test/results/clientpositive/llap/llap_acid_fast.q.out index 4a7297d9073496772a32f808bbe4e9d0bf94fcb9..fc9ccd5a12edf83c9104ecaa6ab5d8ac3771b4fc 100644 --- ql/src/test/results/clientpositive/llap/llap_acid_fast.q.out +++ ql/src/test/results/clientpositive/llap/llap_acid_fast.q.out @@ -242,12 +242,14 @@ STAGE PLANS: native: true projectedOutputColumnNums: [5, 2, 3, 4] Reduce Sink Vectorization: + bucketColumnNums: [5] className: VectorReduceSinkObjectHashOperator - keyColumnNums: [5] + keyColumnNums: [4, 6, 5] + keyExpressions: ConstantVectorExpression(val _bucket_number) -> 6:string native: true nativeConditionsMet: hive.vectorized.execution.reducesink.new.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true, No PTF TopN IS true, No DISTINCT columns IS true, BinarySortableSerDe for keys IS true, LazyBinarySerDe for values IS true - partitionColumnNums: [6] - valueColumnNums: [2, 3, 4] + partitionColumnNums: [4] + valueColumnNums: [2, 3] Execution mode: vectorized, llap LLAP IO: may be used (ACID table) Map Vectorization: @@ -266,28 +268,28 @@ STAGE PLANS: neededVirtualColumns: [ROWID] partitionColumnCount: 1 partitionColumns: csmallint:smallint - scratchColumnTypeNames: [bigint] + scratchColumnTypeNames: [string] Reducer 2 Execution mode: vectorized, llap Reduce Vectorization: enabled: true enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true, hive.execution.engine tez IN [tez, spark] IS true - reduceColumnNullOrder: a - reduceColumnSortOrder: + + reduceColumnNullOrder: aaa + reduceColumnSortOrder: +++ allNative: false usesVectorUDFAdaptor: false vectorized: true rowBatchContext: - dataColumnCount: 4 - dataColumns: KEY.reducesinkkey0:struct, VALUE._col1:float, VALUE._col2:double, VALUE._col3:smallint + dataColumnCount: 5 + dataColumns: KEY._col5:smallint, KEY.'_bucket_number':string, KEY._col0:struct, VALUE._col3:float, VALUE._col4:double partitionColumnCount: 0 scratchColumnTypeNames: [bigint, bigint] Reduce Operator Tree: Select Vectorization: className: VectorSelectOperator native: true - projectedOutputColumnNums: [0, 4, 5, 1, 2, 3] - selectExpressions: ConstantVectorExpression(val 1) -> 4:int, ConstantVectorExpression(val 2) -> 5:bigint + projectedOutputColumnNums: [2, 5, 6, 3, 4, 0, 1] + selectExpressions: ConstantVectorExpression(val 1) -> 5:int, ConstantVectorExpression(val 2) -> 6:bigint File Sink Vectorization: className: VectorFileSinkOperator native: false @@ -420,7 +422,7 @@ POSTHOOK: Input: default@orc_llap_acid_fast@csmallint=3 #### A masked pattern was here #### -285355633 1 -1241163445 -109813638 1 -58941842 -1 1 2 +1 1 1 164554497 1 1161977292 199879534 1 123351087 246423894 1 -1645852809 diff --git ql/src/test/results/clientpositive/llap/llap_smb.q.out ql/src/test/results/clientpositive/llap/llap_smb.q.out index c3047da75b70aea0f2f22e0d64577f0e039daac0..14938d767b9acfad5b4b2c9d1e080dd5266a4ea4 100644 --- ql/src/test/results/clientpositive/llap/llap_smb.q.out +++ ql/src/test/results/clientpositive/llap/llap_smb.q.out @@ -177,7 +177,7 @@ POSTHOOK: Input: default@orc_b 2000 3 6149 2000 5 5720 2000 9 8151 -2001 0 6721 +2001 0 6435 2001 1 7493 2001 2 6578 2001 4 7865 @@ -187,7 +187,7 @@ POSTHOOK: Input: default@orc_b 2000 8 9438 2001 6 5577 2001 7 6149 -2000 0 6721 +2000 0 6435 2000 4 7865 2000 6 5577 2001 3 6149 @@ -223,10 +223,10 @@ STAGE PLANS: TableScan alias: a filterExpr: id is not null (type: boolean) - Statistics: Num rows: 5000 Data size: 158008 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 4998 Data size: 157952 Basic stats: COMPLETE Column stats: PARTIAL Filter Operator predicate: id is not null (type: boolean) - Statistics: Num rows: 4750 Data size: 74104 Basic stats: COMPLETE Column stats: PARTIAL + Statistics: Num rows: 4749 Data size: 74096 Basic stats: COMPLETE Column stats: PARTIAL Merge Join Operator condition map: Inner Join 0 to 1 @@ -234,18 +234,18 @@ STAGE PLANS: 0 id (type: bigint) 1 id (type: bigint) outputColumnNames: _col2, _col3 - Statistics: Num rows: 5225 Data size: 81514 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 5223 Data size: 81505 Basic stats: COMPLETE Column stats: NONE Group By Operator aggregations: count() keys: _col2 (type: int), _col3 (type: smallint) mode: hash outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 5225 Data size: 81514 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 5223 Data size: 81505 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: int), _col1 (type: smallint) sort order: ++ Map-reduce partition columns: _col0 (type: int), _col1 (type: smallint) - Statistics: Num rows: 5225 Data size: 81514 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 5223 Data size: 81505 Basic stats: COMPLETE Column stats: NONE value expressions: _col2 (type: bigint) Execution mode: llap Reducer 2 @@ -256,10 +256,10 @@ STAGE PLANS: keys: KEY._col0 (type: int), KEY._col1 (type: smallint) mode: mergepartial outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 2612 Data size: 40749 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2611 Data size: 40744 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 2612 Data size: 40749 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2611 Data size: 40744 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.SequenceFileInputFormat output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat @@ -326,7 +326,7 @@ POSTHOOK: Input: default@orc_b 2000 3 6149 2000 5 5720 2000 9 8151 -2001 0 6721 +2001 0 6435 2001 1 7493 2001 2 6578 2001 4 7865 @@ -336,7 +336,7 @@ POSTHOOK: Input: default@orc_b 2000 8 9438 2001 6 5577 2001 7 6149 -2000 0 6721 +2000 0 6435 2000 4 7865 2000 6 5577 2001 3 6149 diff --git ql/src/test/results/clientpositive/llap/load_dyn_part1.q.out ql/src/test/results/clientpositive/llap/load_dyn_part1.q.out index 2b89d24027719fa3e10cb3049fd75ddf83001f15..6ed0835ab54e4b99db8379a2743d5422f1132a7a 100644 --- ql/src/test/results/clientpositive/llap/load_dyn_part1.q.out +++ ql/src/test/results/clientpositive/llap/load_dyn_part1.q.out @@ -63,6 +63,10 @@ STAGE PLANS: Stage: Stage-2 Tez #### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) + Reducer 3 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### Vertices: Map 1 Map Operator Tree: @@ -76,14 +80,12 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part1 + value expressions: _col0 (type: string), _col1 (type: string) Filter Operator predicate: (ds > '2008-04-08') (type: boolean) Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE @@ -91,16 +93,46 @@ STAGE PLANS: expressions: key (type: string), value (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 666 Data size: 241092 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 666 Data size: 241092 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part2 + value expressions: _col0 (type: string), _col1 (type: string) Execution mode: llap LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 666 Data size: 363636 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part1 + Reducer 3 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 666 Data size: 241092 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 666 Data size: 241092 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part2 Stage: Stage-3 Dependency Collection diff --git ql/src/test/results/clientpositive/llap/load_dyn_part3.q.out ql/src/test/results/clientpositive/llap/load_dyn_part3.q.out index 5230b597c04685073467222a95696600448cf0e3..266f0114c1202cf72125dd39a83d21c6f2a7a585 100644 --- ql/src/test/results/clientpositive/llap/load_dyn_part3.q.out +++ ql/src/test/results/clientpositive/llap/load_dyn_part3.q.out @@ -49,6 +49,9 @@ STAGE PLANS: Stage: Stage-1 Tez #### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### Vertices: Map 1 Map Operator Tree: @@ -59,16 +62,30 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part3 + value expressions: _col0 (type: string), _col1 (type: string) Execution mode: llap LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2000 Data size: 1092000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part3 Stage: Stage-2 Dependency Collection diff --git ql/src/test/results/clientpositive/llap/load_dyn_part5.q.out ql/src/test/results/clientpositive/llap/load_dyn_part5.q.out index 8b98b8ed9cbcb20eca8dce51c1d378edf3ad695e..1bee446a4bffa30142185ab10a58a9e3f2f7d161 100644 --- ql/src/test/results/clientpositive/llap/load_dyn_part5.q.out +++ ql/src/test/results/clientpositive/llap/load_dyn_part5.q.out @@ -36,6 +36,9 @@ STAGE PLANS: Stage: Stage-1 Tez #### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### Vertices: Map 1 Map Operator Tree: @@ -46,16 +49,30 @@ STAGE PLANS: expressions: key (type: string), value (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col1 (type: string) + sort order: + + Map-reduce partition columns: _col1 (type: string) Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part5 + value expressions: _col0 (type: string) Execution mode: llap LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), KEY._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 500 Data size: 89000 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part5 Stage: Stage-2 Dependency Collection diff --git ql/src/test/results/clientpositive/llap/orc_merge2.q.out ql/src/test/results/clientpositive/llap/orc_merge2.q.out index 53a64248a340cfebaa0311b7476d9dd5f6d22d3b..dff4f26e6348e06ce17f4f7c40c0683aa57df97a 100644 --- ql/src/test/results/clientpositive/llap/orc_merge2.q.out +++ ql/src/test/results/clientpositive/llap/orc_merge2.q.out @@ -34,6 +34,9 @@ STAGE PLANS: Stage: Stage-1 Tez #### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### Vertices: Map 1 Map Operator Tree: @@ -44,16 +47,30 @@ STAGE PLANS: expressions: UDFToInteger(key) (type: int), value (type: string), (hash(key) pmod 10) (type: int), (hash(value) pmod 10) (type: int) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: int), _col3 (type: int) + sort order: ++ + Map-reduce partition columns: _col2 (type: int), _col3 (type: int) Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat - serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde - name: default.orcfile_merge2a + value expressions: _col0 (type: int), _col1 (type: string) Execution mode: llap LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: string), KEY._col2 (type: int), KEY._col3 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 500 Data size: 51500 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat + output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat + serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde + name: default.orcfile_merge2a Stage: Stage-2 Dependency Collection diff --git ql/src/test/results/clientpositive/llap/ppd_union_view.q.out ql/src/test/results/clientpositive/llap/ppd_union_view.q.out index 6eb29208c1faf33eaef11e1d3a3142a160c07227..52f2ce556b0f609f10a8f3f38dba907c72b13a60 100644 --- ql/src/test/results/clientpositive/llap/ppd_union_view.q.out +++ ql/src/test/results/clientpositive/llap/ppd_union_view.q.out @@ -258,6 +258,8 @@ STAGE PLANS: columns.comments columns.types string:string #### A masked pattern was here #### + insideView TRUE +#### A masked pattern was here #### name default.t1_old partition_columns ds partition_columns.types string @@ -337,6 +339,8 @@ STAGE PLANS: columns.comments columns.types string:string #### A masked pattern was here #### + insideView TRUE +#### A masked pattern was here #### name default.t1_mapping partition_columns ds partition_columns.types string @@ -536,6 +540,8 @@ STAGE PLANS: columns.comments columns.types string:string #### A masked pattern was here #### + insideView TRUE +#### A masked pattern was here #### name default.t1_new partition_columns ds partition_columns.types string diff --git ql/src/test/results/clientpositive/llap/schema_evol_orc_acid_part_update.q.out ql/src/test/results/clientpositive/llap/schema_evol_orc_acid_part_update.q.out index 2bcb6d5b6e7732ff1b9fca427d45e5c2889651b6..b3e178d33d650e697253a27a77e739962b870e3e 100644 --- ql/src/test/results/clientpositive/llap/schema_evol_orc_acid_part_update.q.out +++ ql/src/test/results/clientpositive/llap/schema_evol_orc_acid_part_update.q.out @@ -146,21 +146,23 @@ POSTHOOK: Input: default@partitioned_update_1@part=1 POSTHOOK: Input: default@partitioned_update_1@part=2 #### A masked pattern was here #### insert_num part a b c d -101 1 -2147483648 original 99 NULL +101 1 -2147483648 original NULL NULL 102 1 2147483647 original 99 NULL +102 1 2147483647 original NULL NULL 103 1 NULL original 99 NULL -104 1 203332 original 99 NULL +104 1 203332 original NULL NULL 105 1 -999992 original 99 NULL 106 2 0 new 99 hangar paralysed companion -107 2 127 new 99 bottom -108 2 NULL new 99 baffling +107 2 127 new 127 bottom +108 2 NULL new NULL baffling +109 2 17 new 17 feather weight 109 2 17 new 99 feather weight 110 2 400 new 99 window 111 1 -57923222 new 99 astonishing -112 1 82153733 new 99 Antarctic peninsula -113 1 40000 new 99 Delaware +112 1 82153733 new 82153733 Antarctic peninsula +113 1 40000 new 40000 Delaware 114 1 -128 new 99 smaller than Scotland -115 1 32768 new 99 major glaciation +115 1 32768 new 32768 major glaciation PREHOOK: query: alter table partitioned_update_1 partition(part=1) compact 'major' PREHOOK: type: ALTERTABLE_COMPACT POSTHOOK: query: alter table partitioned_update_1 partition(part=1) compact 'major' @@ -182,21 +184,23 @@ POSTHOOK: Input: default@partitioned_update_1@part=1 POSTHOOK: Input: default@partitioned_update_1@part=2 #### A masked pattern was here #### insert_num part a b c d -101 1 -2147483648 original 99 NULL +101 1 -2147483648 original NULL NULL 102 1 2147483647 original 99 NULL +102 1 2147483647 original NULL NULL 103 1 NULL original 99 NULL -104 1 203332 original 99 NULL +104 1 203332 original NULL NULL 105 1 -999992 original 99 NULL 106 2 0 new 99 hangar paralysed companion -107 2 127 new 99 bottom -108 2 NULL new 99 baffling +107 2 127 new 127 bottom +108 2 NULL new NULL baffling +109 2 17 new 17 feather weight 109 2 17 new 99 feather weight 110 2 400 new 99 window 111 1 -57923222 new 99 astonishing -112 1 82153733 new 99 Antarctic peninsula -113 1 40000 new 99 Delaware +112 1 82153733 new 82153733 Antarctic peninsula +113 1 40000 new 40000 Delaware 114 1 -128 new 99 smaller than Scotland -115 1 32768 new 99 major glaciation +115 1 32768 new 32768 major glaciation PREHOOK: query: DROP TABLE partitioned_update_1 PREHOOK: type: DROPTABLE PREHOOK: Input: default@partitioned_update_1 @@ -494,6 +498,7 @@ insert_num part a b c d 111 1 -57923222 new -57923222 astonishing 112 1 82153733 new 82153733 Antarctic peninsula 113 1 40000 new 40000 Delaware +115 1 32768 new 32768 major glaciation PREHOOK: query: alter table partitioned_delete_2 partition(part=1) compact 'major' PREHOOK: type: ALTERTABLE_COMPACT POSTHOOK: query: alter table partitioned_delete_2 partition(part=1) compact 'major' @@ -527,6 +532,7 @@ insert_num part a b c d 111 1 -57923222 new -57923222 astonishing 112 1 82153733 new 82153733 Antarctic peninsula 113 1 40000 new 40000 Delaware +115 1 32768 new 32768 major glaciation PREHOOK: query: DROP TABLE partitioned_delete_2 PREHOOK: type: DROPTABLE PREHOOK: Input: default@partitioned_delete_2 diff --git ql/src/test/results/clientpositive/llap/schema_evol_orc_acidvec_part_update.q.out ql/src/test/results/clientpositive/llap/schema_evol_orc_acidvec_part_update.q.out index f56ec4cb70dc69ecaae4f12fcc8e83b758674875..622ae0671d3b0e7b4d1f2311c51bbdabe2247957 100644 --- ql/src/test/results/clientpositive/llap/schema_evol_orc_acidvec_part_update.q.out +++ ql/src/test/results/clientpositive/llap/schema_evol_orc_acidvec_part_update.q.out @@ -146,21 +146,21 @@ POSTHOOK: Input: default@partitioned_update_1@part=1 POSTHOOK: Input: default@partitioned_update_1@part=2 #### A masked pattern was here #### insert_num part a b c d -101 1 -2147483648 original 99 NULL -102 1 2147483647 original 99 NULL -103 1 NULL original 99 NULL -104 1 203332 original 99 NULL -105 1 -999992 original 99 NULL -106 2 0 new 99 hangar paralysed companion -107 2 127 new 99 bottom -108 2 NULL new 99 baffling -109 2 17 new 99 feather weight -110 2 400 new 99 window -111 1 -57923222 new 99 astonishing -112 1 82153733 new 99 Antarctic peninsula -113 1 40000 new 99 Delaware -114 1 -128 new 99 smaller than Scotland -115 1 32768 new 99 major glaciation +101 1 -2147483648 original NULL NULL +102 1 2147483647 original NULL NULL +103 1 NULL original NULL NULL +104 1 203332 original NULL NULL +105 1 -999992 original NULL NULL +106 2 0 new 0 hangar paralysed companion +107 2 127 new 127 bottom +108 2 NULL new NULL baffling +109 2 17 new 17 feather weight +110 2 400 new 400 window +111 1 -57923222 new -57923222 astonishing +112 1 82153733 new 82153733 Antarctic peninsula +113 1 40000 new 40000 Delaware +114 1 -128 new -128 smaller than Scotland +115 1 32768 new 32768 major glaciation PREHOOK: query: alter table partitioned_update_1 partition(part=1) compact 'major' PREHOOK: type: ALTERTABLE_COMPACT POSTHOOK: query: alter table partitioned_update_1 partition(part=1) compact 'major' @@ -182,21 +182,21 @@ POSTHOOK: Input: default@partitioned_update_1@part=1 POSTHOOK: Input: default@partitioned_update_1@part=2 #### A masked pattern was here #### insert_num part a b c d -101 1 -2147483648 original 99 NULL -102 1 2147483647 original 99 NULL -103 1 NULL original 99 NULL -104 1 203332 original 99 NULL -105 1 -999992 original 99 NULL -106 2 0 new 99 hangar paralysed companion -107 2 127 new 99 bottom -108 2 NULL new 99 baffling -109 2 17 new 99 feather weight -110 2 400 new 99 window -111 1 -57923222 new 99 astonishing -112 1 82153733 new 99 Antarctic peninsula -113 1 40000 new 99 Delaware -114 1 -128 new 99 smaller than Scotland -115 1 32768 new 99 major glaciation +101 1 -2147483648 original NULL NULL +102 1 2147483647 original NULL NULL +103 1 NULL original NULL NULL +104 1 203332 original NULL NULL +105 1 -999992 original NULL NULL +106 2 0 new 0 hangar paralysed companion +107 2 127 new 127 bottom +108 2 NULL new NULL baffling +109 2 17 new 17 feather weight +110 2 400 new 400 window +111 1 -57923222 new -57923222 astonishing +112 1 82153733 new 82153733 Antarctic peninsula +113 1 40000 new 40000 Delaware +114 1 -128 new -128 smaller than Scotland +115 1 32768 new 32768 major glaciation PREHOOK: query: DROP TABLE partitioned_update_1 PREHOOK: type: DROPTABLE PREHOOK: Input: default@partitioned_update_1 @@ -318,8 +318,11 @@ POSTHOOK: Input: default@partitioned_delete_1@part=2 #### A masked pattern was here #### insert_num part a b c d 101 1 -2147483648 original NULL NULL +102 1 2147483647 original NULL NULL 103 1 NULL original NULL NULL +104 1 203332 original NULL NULL 105 1 -999992 original NULL NULL +106 2 0 new 0 hangar paralysed companion 107 2 127 new 127 bottom 108 2 NULL new NULL baffling 109 2 17 new 17 feather weight @@ -351,8 +354,11 @@ POSTHOOK: Input: default@partitioned_delete_1@part=2 #### A masked pattern was here #### insert_num part a b c d 101 1 -2147483648 original NULL NULL +102 1 2147483647 original NULL NULL 103 1 NULL original NULL NULL +104 1 203332 original NULL NULL 105 1 -999992 original NULL NULL +106 2 0 new 0 hangar paralysed companion 107 2 127 new 127 bottom 108 2 NULL new NULL baffling 109 2 17 new 17 feather weight @@ -489,11 +495,14 @@ insert_num part a b c d 105 1 -999992 original NULL NULL 106 2 0 new 0 hangar paralysed companion 107 2 127 new 127 bottom +108 2 NULL new NULL baffling 109 2 17 new 17 feather weight 110 2 400 new 400 window 111 1 -57923222 new -57923222 astonishing 112 1 82153733 new 82153733 Antarctic peninsula 113 1 40000 new 40000 Delaware +114 1 -128 new -128 smaller than Scotland +115 1 32768 new 32768 major glaciation PREHOOK: query: alter table partitioned_delete_2 partition(part=1) compact 'major' PREHOOK: type: ALTERTABLE_COMPACT POSTHOOK: query: alter table partitioned_delete_2 partition(part=1) compact 'major' @@ -522,11 +531,14 @@ insert_num part a b c d 105 1 -999992 original NULL NULL 106 2 0 new 0 hangar paralysed companion 107 2 127 new 127 bottom +108 2 NULL new NULL baffling 109 2 17 new 17 feather weight 110 2 400 new 400 window 111 1 -57923222 new -57923222 astonishing 112 1 82153733 new 82153733 Antarctic peninsula 113 1 40000 new 40000 Delaware +114 1 -128 new -128 smaller than Scotland +115 1 32768 new 32768 major glaciation PREHOOK: query: DROP TABLE partitioned_delete_2 PREHOOK: type: DROPTABLE PREHOOK: Input: default@partitioned_delete_2 diff --git ql/src/test/results/clientpositive/llap/sysdb.q.out ql/src/test/results/clientpositive/llap/sysdb.q.out index 981363653115ca08dbe51360a83fea21990b87b8..94199521512445ae65a6de2f3d58b7e18689d8fd 100644 --- ql/src/test/results/clientpositive/llap/sysdb.q.out +++ ql/src/test/results/clientpositive/llap/sysdb.q.out @@ -2190,18 +2190,20 @@ POSTHOOK: Lineage: PARTITION_STATS_VIEW.transient_last_ddl_time EXPRESSION [(par PREHOOK: query: CREATE TABLE IF NOT EXISTS `WM_RESOURCEPLANS` ( `NAME` string, `STATUS` string, - `QUERY_PARALLELISM` int + `QUERY_PARALLELISM` int, + `DEFAULT_POOL_PATH` string ) STORED BY 'org.apache.hive.storage.jdbc.JdbcStorageHandler' TBLPROPERTIES ( "hive.sql.database.type" = "METASTORE", "hive.sql.query" = "SELECT - \"NAME\", + \"WM_RESOURCEPLAN\".\"NAME\", \"STATUS\", - \"QUERY_PARALLELISM\" + \"WM_RESOURCEPLAN\".\"QUERY_PARALLELISM\", + \"WM_POOL\".\"PATH\" FROM - \"WM_RESOURCEPLAN\"" + \"WM_RESOURCEPLAN\" LEFT OUTER JOIN \"WM_POOL\" ON \"WM_RESOURCEPLAN\".\"DEFAULT_POOL_ID\" = \"WM_POOL\".\"POOL_ID\"" ) PREHOOK: type: CREATETABLE PREHOOK: Output: SYS@WM_RESOURCEPLANS @@ -2209,22 +2211,74 @@ PREHOOK: Output: database:sys POSTHOOK: query: CREATE TABLE IF NOT EXISTS `WM_RESOURCEPLANS` ( `NAME` string, `STATUS` string, - `QUERY_PARALLELISM` int + `QUERY_PARALLELISM` int, + `DEFAULT_POOL_PATH` string ) STORED BY 'org.apache.hive.storage.jdbc.JdbcStorageHandler' TBLPROPERTIES ( "hive.sql.database.type" = "METASTORE", "hive.sql.query" = "SELECT - \"NAME\", + \"WM_RESOURCEPLAN\".\"NAME\", \"STATUS\", - \"QUERY_PARALLELISM\" + \"WM_RESOURCEPLAN\".\"QUERY_PARALLELISM\", + \"WM_POOL\".\"PATH\" FROM - \"WM_RESOURCEPLAN\"" + \"WM_RESOURCEPLAN\" LEFT OUTER JOIN \"WM_POOL\" ON \"WM_RESOURCEPLAN\".\"DEFAULT_POOL_ID\" = \"WM_POOL\".\"POOL_ID\"" ) POSTHOOK: type: CREATETABLE POSTHOOK: Output: SYS@WM_RESOURCEPLANS POSTHOOK: Output: database:sys +PREHOOK: query: CREATE TABLE IF NOT EXISTS `WM_TRIGGERS` ( + `RP_NAME` string, + `NAME` string, + `TRIGGER_EXPRESSION` string, + `ACTION_EXPRESSION` string +) +STORED BY 'org.apache.hive.storage.jdbc.JdbcStorageHandler' +TBLPROPERTIES ( +"hive.sql.database.type" = "METASTORE", +"hive.sql.query" = +"SELECT + r.NAME RP_NAME, + t.NAME NAME, + TRIGGER_EXPRESSION, + ACTION_EXPRESSION +FROM + WM_TRIGGER t +JOIN + WM_RESOURCEPLAN r +ON + t.RP_ID = r.RP_ID" +) +PREHOOK: type: CREATETABLE +PREHOOK: Output: SYS@WM_TRIGGERS +PREHOOK: Output: database:sys +POSTHOOK: query: CREATE TABLE IF NOT EXISTS `WM_TRIGGERS` ( + `RP_NAME` string, + `NAME` string, + `TRIGGER_EXPRESSION` string, + `ACTION_EXPRESSION` string +) +STORED BY 'org.apache.hive.storage.jdbc.JdbcStorageHandler' +TBLPROPERTIES ( +"hive.sql.database.type" = "METASTORE", +"hive.sql.query" = +"SELECT + r.NAME RP_NAME, + t.NAME NAME, + TRIGGER_EXPRESSION, + ACTION_EXPRESSION +FROM + WM_TRIGGER t +JOIN + WM_RESOURCEPLAN r +ON + t.RP_ID = r.RP_ID" +) +POSTHOOK: type: CREATETABLE +POSTHOOK: Output: SYS@WM_TRIGGERS +POSTHOOK: Output: database:sys PREHOOK: query: DROP DATABASE IF EXISTS INFORMATION_SCHEMA PREHOOK: type: DROPDATABASE POSTHOOK: query: DROP DATABASE IF EXISTS INFORMATION_SCHEMA @@ -3005,7 +3059,7 @@ POSTHOOK: query: select count(*) from cds POSTHOOK: type: QUERY POSTHOOK: Input: sys@cds #### A masked pattern was here #### -66 +67 PREHOOK: query: select column_name, type_name, integer_idx from columns_v2 order by column_name, integer_idx limit 5 PREHOOK: type: QUERY PREHOOK: Input: sys@columns_v2 @@ -3015,10 +3069,10 @@ POSTHOOK: type: QUERY POSTHOOK: Input: sys@columns_v2 #### A masked pattern was here #### a decimal(10,2) 0 +action_expression string 3 add_time int 1 aint int 0 astring string 1 -attributes map,map>>>> 6 PREHOOK: query: select param_key, param_value from database_params order by param_key, param_value limit 5 PREHOOK: type: QUERY PREHOOK: Input: sys@database_params @@ -3159,7 +3213,7 @@ POSTHOOK: query: select count(*) from sds POSTHOOK: type: QUERY POSTHOOK: Input: sys@sds #### A masked pattern was here #### -72 +73 PREHOOK: query: select param_key, param_value from sd_params order by param_key, param_value limit 5 PREHOOK: type: QUERY PREHOOK: Input: sys@sd_params @@ -3269,10 +3323,10 @@ POSTHOOK: type: QUERY POSTHOOK: Input: sys@table_params #### A masked pattern was here #### COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true","c":"true","d":"true","e":"true","f":"true","g":"true"}} +COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"action_expression":"true","name":"true","rp_name":"true","trigger_expression":"true"}} COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"add_time":"true","grant_option":"true","grantor":"true","grantor_type":"true","principal_name":"true","principal_type":"true","role_grant_id":"true","role_id":"true"}} COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"avg_col_len":"true","big_decimal_high_value":"true","big_decimal_low_value":"true","column_name":"true","column_type":"true","cs_id":"true","db_name":"true","double_high_value":"true","double_low_value":"true","last_analyzed":"true","long_high_value":"true","long_low_value":"true","max_col_len":"true","num_distincts":"true","num_falses":"true","num_nulls":"true","num_trues":"true","part_id":"true","partition_name":"true","table_name":"true"}} COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"avg_col_len":"true","big_decimal_high_value":"true","big_decimal_low_value":"true","column_name":"true","column_type":"true","cs_id":"true","db_name":"true","double_high_value":"true","double_low_value":"true","last_analyzed":"true","long_high_value":"true","long_low_value":"true","max_col_len":"true","num_distincts":"true","num_falses":"true","num_nulls":"true","num_trues":"true","table_name":"true","tbl_id":"true"}} -COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"bucket_col_name":"true","integer_idx":"true","sd_id":"true"}} PREHOOK: query: select tbl_name from tbls order by tbl_name limit 5 PREHOOK: type: QUERY PREHOOK: Input: sys@tbls @@ -3384,11 +3438,11 @@ POSTHOOK: type: QUERY POSTHOOK: Input: sys@table_params POSTHOOK: Input: sys@table_stats_view #### A masked pattern was here #### -{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} 0 0 0 0 -{"BASIC_STATS":"true","COLUMN_STATS":{"name":"true","query_parallelism":"true","status":"true"}} 0 0 0 0 +{"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}} 0 0 0 0 +{"BASIC_STATS":"true","COLUMN_STATS":{"default_pool_path":"true","name":"true","query_parallelism":"true","status":"true"}} 0 0 0 0 {"BASIC_STATS":"true","COLUMN_STATS":{"next_val":"true","sequence_name":"true"}} 0 0 0 0 {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} 0 0 0 0 -#### A masked pattern was here #### +{"BASIC_STATS":"true","COLUMN_STATS":{"integer_idx":"true","pkey_comment":"true","pkey_name":"true","pkey_type":"true","tbl_id":"true"}} 0 0 0 0 PREHOOK: query: select COLUMN_STATS_ACCURATE, NUM_FILES, NUM_ROWS, RAW_DATA_SIZE, TOTAL_SIZE FROM PARTITION_STATS_VIEW where COLUMN_STATS_ACCURATE is not null order by NUM_FILES, NUM_ROWS, RAW_DATA_SIZE limit 5 PREHOOK: type: QUERY PREHOOK: Input: sys@partition_params @@ -3591,6 +3645,7 @@ default sys tbl_privs BASE_TABLE NULL NULL NULL NULL NULL YES NO NULL default sys tbls BASE_TABLE NULL NULL NULL NULL NULL YES NO NULL default sys version BASE_TABLE NULL NULL NULL NULL NULL YES NO NULL default sys wm_resourceplans BASE_TABLE NULL NULL NULL NULL NULL YES NO NULL +default sys wm_triggers BASE_TABLE NULL NULL NULL NULL NULL YES NO NULL PREHOOK: query: select * from TABLE_PRIVILEGES order by GRANTOR, GRANTEE, TABLE_SCHEMA, TABLE_NAME, PRIVILEGE_TYPE limit 10 PREHOOK: type: QUERY PREHOOK: Input: information_schema@table_privileges @@ -3643,13 +3698,13 @@ default default alltypesorc ctimestamp1 8 NULL YES timestamp NULL NULL NULL NULL default default alltypesorc ctimestamp2 9 NULL YES timestamp NULL NULL NULL NULL NULL 9 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 11 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES timestamp NULL NULL default default alltypesorc cboolean1 10 NULL YES boolean NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 11 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES boolean NULL NULL default default alltypesorc cboolean2 11 NULL YES boolean NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 11 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES boolean NULL NULL -default default moretypes a 0 NULL YES decimal(10,2) NULL NULL 10 10 2 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 27 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES decimal(10,2) 10 10 -default default moretypes b 1 NULL YES tinyint NULL NULL 3 10 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 27 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES tinyint 3 10 -default default moretypes c 2 NULL YES smallint NULL NULL 5 10 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 27 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES smallint 5 10 -default default moretypes d 3 NULL YES int NULL NULL 10 10 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 27 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES int 10 10 -default default moretypes e 4 NULL YES bigint NULL NULL 19 10 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 27 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES bigint 19 10 -default default moretypes f 5 NULL YES varchar(10) 10 10 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 27 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES varchar(10) NULL NULL -default default moretypes g 6 NULL YES char(3) 3 3 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 27 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES char(3) NULL NULL +default default moretypes a 0 NULL YES decimal(10,2) NULL NULL 10 10 2 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 49 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES decimal(10,2) 10 10 +default default moretypes b 1 NULL YES tinyint NULL NULL 3 10 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 49 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES tinyint 3 10 +default default moretypes c 2 NULL YES smallint NULL NULL 5 10 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 49 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES smallint 5 10 +default default moretypes d 3 NULL YES int NULL NULL 10 10 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 49 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES int 10 10 +default default moretypes e 4 NULL YES bigint NULL NULL 19 10 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 49 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES bigint 19 10 +default default moretypes f 5 NULL YES varchar(10) 10 10 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 49 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES varchar(10) NULL NULL +default default moretypes g 6 NULL YES char(3) 3 3 NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL NULL 49 NO NO NULL NULL NULL NULL NULL NULL NEVER NULL NO NO NULL YES char(3) NULL NULL PREHOOK: query: select * from COLUMN_PRIVILEGES order by GRANTOR, GRANTEE, TABLE_SCHEMA, TABLE_NAME, COLUMN_NAME limit 10 PREHOOK: type: QUERY PREHOOK: Input: information_schema@column_privileges diff --git ql/src/test/results/clientpositive/llap/tez_dml.q.out ql/src/test/results/clientpositive/llap/tez_dml.q.out index adccaae2aa0944a08c388aaebd1a852cb294b8e2..ae33c648bb5be5631890c3016f6bc8e8cec71aa3 100644 --- ql/src/test/results/clientpositive/llap/tez_dml.q.out +++ ql/src/test/results/clientpositive/llap/tez_dml.q.out @@ -444,6 +444,9 @@ STAGE PLANS: Stage: Stage-1 Tez #### A masked pattern was here #### + Edges: + Reducer 2 <- Map 1 (SIMPLE_EDGE) +#### A masked pattern was here #### Vertices: Map 1 Map Operator Tree: @@ -454,16 +457,30 @@ STAGE PLANS: expressions: value (type: string), cnt (type: bigint) outputColumnNames: _col0, _col1 Statistics: Num rows: 309 Data size: 56640 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col1 (type: bigint) + sort order: + + Map-reduce partition columns: _col1 (type: bigint) Statistics: Num rows: 309 Data size: 56640 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.tmp_src_part + value expressions: _col0 (type: string) Execution mode: llap LLAP IO: no inputs + Reducer 2 + Execution mode: llap + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), KEY._col1 (type: bigint) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 309 Data size: 56640 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 309 Data size: 56640 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.tmp_src_part Stage: Stage-2 Dependency Collection diff --git ql/src/test/results/clientpositive/llap/update_all_partitioned.q.out ql/src/test/results/clientpositive/llap/update_all_partitioned.q.out index 4ee1e2e4cf5d5b056a6e5b0dc38fe39b35d26c92..7301f5990215011dda6dcd0bce2d33e67486a488 100644 --- ql/src/test/results/clientpositive/llap/update_all_partitioned.q.out +++ ql/src/test/results/clientpositive/llap/update_all_partitioned.q.out @@ -84,23 +84,29 @@ POSTHOOK: Input: default@acid_uap POSTHOOK: Input: default@acid_uap@ds=today POSTHOOK: Input: default@acid_uap@ds=tomorrow #### A masked pattern was here #### --1073279343 fred today --1073051226 fred today +-1073279343 oj1YrV5Wa today +-1073051226 A34p7oRr2WvUJNf today +-1072910839 0iqrc5 today -1072910839 fred today --1072081801 fred today +-1072081801 dPkN74F7 today -1072076362 fred today --1071480828 fred today +-1071480828 aw724t8c5558x2xneC624 today +-1071363017 Anj0oF today -1071363017 fred today --1070883071 fred today +-1070883071 0ruyd6Y50JpdGRf6HqD today -1070551679 fred today --1069736047 fred today -762 fred tomorrow -762 fred tomorrow +-1070551679 iUR3Q today +-1069736047 k17Am8uPHWk02cEf1jet today +762 40ks5556SV tomorrow +762 a10E76jX35YwquKCTA tomorrow 762 fred tomorrow 762 fred tomorrow 762 fred tomorrow -6981 fred tomorrow -6981 fred tomorrow +6981 NULL tomorrow +6981 1FNNhmiFLGw425NA13g tomorrow +6981 4KhrrQ0nJ7bMNTvhSCA tomorrow +6981 K630vaVf tomorrow +6981 Y5x3JuI3M8jngv5N tomorrow 6981 fred tomorrow 6981 fred tomorrow 6981 fred tomorrow diff --git ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out index 1fb6e9d48cc5282d98fe889c607e7b3fa215ed12..2d53c8d4b008ba40cfa9aa0de1d4fe11f3de028d 100644 --- ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out +++ ql/src/test/results/clientpositive/llap/vector_partitioned_date_time.q.out @@ -683,101 +683,92 @@ POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-30 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-31 #### A masked pattern was here #### Baltimore New York 2010-10-26 07:00:00 -22.0 1064 2010-10-26 -Baltimore New York 2010-10-26 07:00:00 123.0 1142 2010-10-26 -Baltimore New York 2010-10-26 07:00:00 90.0 1599 2010-10-26 -Chicago New York 2010-10-26 07:00:00 12.0 361 2010-10-26 -Chicago New York 2010-10-26 07:00:00 0.0 897 2010-10-26 -Chicago New York 2010-10-26 07:00:00 29.0 1531 2010-10-26 -Chicago New York 2010-10-26 07:00:00 -17.0 1610 2010-10-26 -Chicago New York 2010-10-26 07:00:00 6.0 3198 2010-10-26 +Cleveland New York 2010-10-26 07:00:00 13.0 3014 2010-10-26 Cleveland New York 2010-10-26 07:00:00 4.0 2630 2010-10-26 Cleveland New York 2010-10-26 07:00:00 -27.0 2646 2010-10-26 Cleveland New York 2010-10-26 07:00:00 -11.0 2662 2010-10-26 -Cleveland New York 2010-10-26 07:00:00 13.0 3014 2010-10-26 Washington New York 2010-10-26 07:00:00 4.0 7291 2010-10-26 -Baltimore New York 2010-10-21 07:00:00 17.0 1064 2010-10-21 -Baltimore New York 2010-10-21 07:00:00 105.0 1142 2010-10-21 -Baltimore New York 2010-10-21 07:00:00 28.0 1599 2010-10-21 -Chicago New York 2010-10-21 07:00:00 142.0 361 2010-10-21 -Chicago New York 2010-10-21 07:00:00 77.0 897 2010-10-21 -Chicago New York 2010-10-21 07:00:00 53.0 1531 2010-10-21 -Chicago New York 2010-10-21 07:00:00 -5.0 1610 2010-10-21 -Chicago New York 2010-10-21 07:00:00 51.0 3198 2010-10-21 -Cleveland New York 2010-10-21 07:00:00 3.0 2630 2010-10-21 -Cleveland New York 2010-10-21 07:00:00 29.0 2646 2010-10-21 -Cleveland New York 2010-10-21 07:00:00 72.0 3014 2010-10-21 -Washington New York 2010-10-21 07:00:00 22.0 7291 2010-10-21 +Chicago New York 2010-10-26 07:00:00 6.0 3198 2010-10-26 +Chicago New York 2010-10-26 07:00:00 -17.0 1610 2010-10-26 +Chicago New York 2010-10-26 07:00:00 29.0 1531 2010-10-26 +Chicago New York 2010-10-26 07:00:00 0.0 897 2010-10-26 +Chicago New York 2010-10-26 07:00:00 12.0 361 2010-10-26 +Baltimore New York 2010-10-26 07:00:00 90.0 1599 2010-10-26 +Baltimore New York 2010-10-26 07:00:00 123.0 1142 2010-10-26 +Chicago New York 2010-10-25 07:00:00 6.0 1610 2010-10-25 +Chicago New York 2010-10-25 07:00:00 -16.0 3198 2010-10-25 Baltimore New York 2010-10-25 07:00:00 -25.0 1064 2010-10-25 Baltimore New York 2010-10-25 07:00:00 92.0 1142 2010-10-25 Baltimore New York 2010-10-25 07:00:00 106.0 1599 2010-10-25 Chicago New York 2010-10-25 07:00:00 31.0 361 2010-10-25 -Chicago New York 2010-10-25 07:00:00 -1.0 897 2010-10-25 -Chicago New York 2010-10-25 07:00:00 43.0 1531 2010-10-25 -Chicago New York 2010-10-25 07:00:00 6.0 1610 2010-10-25 -Chicago New York 2010-10-25 07:00:00 -16.0 3198 2010-10-25 +Cleveland New York 2010-10-25 07:00:00 42.0 3014 2010-10-25 Cleveland New York 2010-10-25 07:00:00 -4.0 2630 2010-10-25 Cleveland New York 2010-10-25 07:00:00 81.0 2646 2010-10-25 -Cleveland New York 2010-10-25 07:00:00 42.0 3014 2010-10-25 Washington New York 2010-10-25 07:00:00 9.0 7291 2010-10-25 -Baltimore New York 2010-10-24 07:00:00 12.0 1599 2010-10-24 -Baltimore New York 2010-10-24 07:00:00 20.0 2571 2010-10-24 -Chicago New York 2010-10-24 07:00:00 10.0 361 2010-10-24 -Chicago New York 2010-10-24 07:00:00 113.0 897 2010-10-24 +Chicago New York 2010-10-25 07:00:00 -1.0 897 2010-10-25 +Chicago New York 2010-10-25 07:00:00 43.0 1531 2010-10-25 +Chicago New York 2010-10-21 07:00:00 142.0 361 2010-10-21 +Chicago New York 2010-10-21 07:00:00 77.0 897 2010-10-21 +Chicago New York 2010-10-21 07:00:00 53.0 1531 2010-10-21 +Chicago New York 2010-10-21 07:00:00 -5.0 1610 2010-10-21 +Chicago New York 2010-10-21 07:00:00 51.0 3198 2010-10-21 +Washington New York 2010-10-21 07:00:00 22.0 7291 2010-10-21 +Baltimore New York 2010-10-21 07:00:00 28.0 1599 2010-10-21 +Cleveland New York 2010-10-21 07:00:00 72.0 3014 2010-10-21 +Cleveland New York 2010-10-21 07:00:00 3.0 2630 2010-10-21 +Cleveland New York 2010-10-21 07:00:00 29.0 2646 2010-10-21 +Baltimore New York 2010-10-21 07:00:00 17.0 1064 2010-10-21 +Baltimore New York 2010-10-21 07:00:00 105.0 1142 2010-10-21 Chicago New York 2010-10-24 07:00:00 -5.0 1531 2010-10-24 -Chicago New York 2010-10-24 07:00:00 -17.0 1610 2010-10-24 -Chicago New York 2010-10-24 07:00:00 -3.0 3198 2010-10-24 +Chicago New York 2010-10-24 07:00:00 113.0 897 2010-10-24 +Chicago New York 2010-10-24 07:00:00 10.0 361 2010-10-24 +Baltimore New York 2010-10-24 07:00:00 20.0 2571 2010-10-24 +Baltimore New York 2010-10-24 07:00:00 12.0 1599 2010-10-24 +Washington New York 2010-10-24 07:00:00 -26.0 7282 2010-10-24 Cleveland New York 2010-10-24 07:00:00 5.0 2254 2010-10-24 -Cleveland New York 2010-10-24 07:00:00 -11.0 2630 2010-10-24 -Cleveland New York 2010-10-24 07:00:00 -20.0 2646 2010-10-24 Cleveland New York 2010-10-24 07:00:00 -9.0 3014 2010-10-24 -Washington New York 2010-10-24 07:00:00 -26.0 7282 2010-10-24 -Baltimore New York 2010-10-23 07:00:00 18.0 272 2010-10-23 -Baltimore New York 2010-10-23 07:00:00 -10.0 1805 2010-10-23 -Baltimore New York 2010-10-23 07:00:00 6.0 3171 2010-10-23 -Chicago New York 2010-10-23 07:00:00 3.0 384 2010-10-23 -Chicago New York 2010-10-23 07:00:00 32.0 426 2010-10-23 -Chicago New York 2010-10-23 07:00:00 1.0 650 2010-10-23 -Chicago New York 2010-10-23 07:00:00 11.0 3085 2010-10-23 -Cleveland New York 2010-10-23 07:00:00 -21.0 2932 2010-10-23 -Washington New York 2010-10-23 07:00:00 -25.0 5832 2010-10-23 -Washington New York 2010-10-23 07:00:00 -21.0 5904 2010-10-23 -Washington New York 2010-10-23 07:00:00 -18.0 5917 2010-10-23 -Washington New York 2010-10-23 07:00:00 -16.0 7274 2010-10-23 +Cleveland New York 2010-10-24 07:00:00 -20.0 2646 2010-10-24 +Cleveland New York 2010-10-24 07:00:00 -11.0 2630 2010-10-24 +Chicago New York 2010-10-24 07:00:00 -3.0 3198 2010-10-24 +Chicago New York 2010-10-24 07:00:00 -17.0 1610 2010-10-24 +Chicago New York 2010-10-29 07:00:00 -18.0 1610 2010-10-29 +Cleveland New York 2010-10-29 07:00:00 -12.0 3014 2010-10-29 +Cleveland New York 2010-10-29 07:00:00 -19.0 2646 2010-10-29 +Cleveland New York 2010-10-29 07:00:00 -4.0 2630 2010-10-29 +Washington New York 2010-10-29 07:00:00 1.0 7291 2010-10-29 Baltimore New York 2010-10-29 07:00:00 -24.0 1064 2010-10-29 Baltimore New York 2010-10-29 07:00:00 21.0 1142 2010-10-29 Baltimore New York 2010-10-29 07:00:00 -2.0 1599 2010-10-29 Chicago New York 2010-10-29 07:00:00 -12.0 361 2010-10-29 Chicago New York 2010-10-29 07:00:00 -11.0 897 2010-10-29 Chicago New York 2010-10-29 07:00:00 15.0 1531 2010-10-29 -Chicago New York 2010-10-29 07:00:00 -18.0 1610 2010-10-29 Chicago New York 2010-10-29 07:00:00 -4.0 3198 2010-10-29 -Cleveland New York 2010-10-29 07:00:00 -4.0 2630 2010-10-29 -Cleveland New York 2010-10-29 07:00:00 -19.0 2646 2010-10-29 -Cleveland New York 2010-10-29 07:00:00 -12.0 3014 2010-10-29 -Washington New York 2010-10-29 07:00:00 1.0 7291 2010-10-29 -Baltimore New York 2010-10-27 07:00:00 -18.0 1064 2010-10-27 -Baltimore New York 2010-10-27 07:00:00 49.0 1142 2010-10-27 -Baltimore New York 2010-10-27 07:00:00 92.0 1599 2010-10-27 -Chicago New York 2010-10-27 07:00:00 148.0 361 2010-10-27 -Chicago New York 2010-10-27 07:00:00 -11.0 897 2010-10-27 -Chicago New York 2010-10-27 07:00:00 70.0 1531 2010-10-27 -Chicago New York 2010-10-27 07:00:00 8.0 1610 2010-10-27 -Chicago New York 2010-10-27 07:00:00 21.0 3198 2010-10-27 -Cleveland New York 2010-10-27 07:00:00 16.0 2630 2010-10-27 -Cleveland New York 2010-10-27 07:00:00 27.0 3014 2010-10-27 -Washington New York 2010-10-27 07:00:00 26.0 7291 2010-10-27 +Chicago New York 2010-10-28 07:00:00 3.0 1610 2010-10-28 Baltimore New York 2010-10-28 07:00:00 -4.0 1064 2010-10-28 Baltimore New York 2010-10-28 07:00:00 -14.0 1142 2010-10-28 Baltimore New York 2010-10-28 07:00:00 -14.0 1599 2010-10-28 Chicago New York 2010-10-28 07:00:00 2.0 361 2010-10-28 Chicago New York 2010-10-28 07:00:00 2.0 897 2010-10-28 Chicago New York 2010-10-28 07:00:00 -11.0 1531 2010-10-28 -Chicago New York 2010-10-28 07:00:00 3.0 1610 2010-10-28 Chicago New York 2010-10-28 07:00:00 -18.0 3198 2010-10-28 Cleveland New York 2010-10-28 07:00:00 3.0 2630 2010-10-28 Cleveland New York 2010-10-28 07:00:00 -6.0 2646 2010-10-28 Cleveland New York 2010-10-28 07:00:00 1.0 3014 2010-10-28 Washington New York 2010-10-28 07:00:00 45.0 7291 2010-10-28 +Cleveland New York 2010-10-27 07:00:00 16.0 2630 2010-10-27 +Chicago New York 2010-10-27 07:00:00 21.0 3198 2010-10-27 +Chicago New York 2010-10-27 07:00:00 8.0 1610 2010-10-27 +Chicago New York 2010-10-27 07:00:00 70.0 1531 2010-10-27 +Chicago New York 2010-10-27 07:00:00 -11.0 897 2010-10-27 +Chicago New York 2010-10-27 07:00:00 148.0 361 2010-10-27 +Baltimore New York 2010-10-27 07:00:00 92.0 1599 2010-10-27 +Baltimore New York 2010-10-27 07:00:00 49.0 1142 2010-10-27 +Baltimore New York 2010-10-27 07:00:00 -18.0 1064 2010-10-27 +Washington New York 2010-10-27 07:00:00 26.0 7291 2010-10-27 +Cleveland New York 2010-10-27 07:00:00 27.0 3014 2010-10-27 +Cleveland New York 2010-10-30 07:00:00 -12.0 2932 2010-10-30 +Washington New York 2010-10-30 07:00:00 -16.0 5917 2010-10-30 +Washington New York 2010-10-30 07:00:00 -27.0 5904 2010-10-30 Baltimore New York 2010-10-30 07:00:00 14.0 272 2010-10-30 Baltimore New York 2010-10-30 07:00:00 -1.0 1805 2010-10-30 Baltimore New York 2010-10-30 07:00:00 5.0 3171 2010-10-30 @@ -786,39 +777,48 @@ Chicago New York 2010-10-30 07:00:00 -10.0 426 2010-10-30 Chicago New York 2010-10-30 07:00:00 -5.0 650 2010-10-30 Chicago New York 2010-10-30 07:00:00 -5.0 3085 2010-10-30 Cleveland New York 2010-10-30 07:00:00 -23.0 2018 2010-10-30 -Cleveland New York 2010-10-30 07:00:00 -12.0 2932 2010-10-30 -Washington New York 2010-10-30 07:00:00 -27.0 5904 2010-10-30 -Washington New York 2010-10-30 07:00:00 -16.0 5917 2010-10-30 -Baltimore New York 2010-10-20 07:00:00 -30.0 1064 2010-10-20 -Baltimore New York 2010-10-20 07:00:00 23.0 1142 2010-10-20 +Baltimore New York 2010-10-23 07:00:00 6.0 3171 2010-10-23 +Chicago New York 2010-10-23 07:00:00 1.0 650 2010-10-23 +Washington New York 2010-10-23 07:00:00 -25.0 5832 2010-10-23 +Baltimore New York 2010-10-23 07:00:00 -10.0 1805 2010-10-23 +Baltimore New York 2010-10-23 07:00:00 18.0 272 2010-10-23 +Washington New York 2010-10-23 07:00:00 -21.0 5904 2010-10-23 +Washington New York 2010-10-23 07:00:00 -18.0 5917 2010-10-23 +Chicago New York 2010-10-23 07:00:00 32.0 426 2010-10-23 +Cleveland New York 2010-10-23 07:00:00 -21.0 2932 2010-10-23 +Chicago New York 2010-10-23 07:00:00 11.0 3085 2010-10-23 +Washington New York 2010-10-23 07:00:00 -16.0 7274 2010-10-23 +Chicago New York 2010-10-23 07:00:00 3.0 384 2010-10-23 +Chicago New York 2010-10-20 07:00:00 24.0 897 2010-10-20 Baltimore New York 2010-10-20 07:00:00 6.0 1599 2010-10-20 +Baltimore New York 2010-10-20 07:00:00 23.0 1142 2010-10-20 +Washington New York 2010-10-20 07:00:00 -2.0 7291 2010-10-20 Chicago New York 2010-10-20 07:00:00 42.0 361 2010-10-20 -Chicago New York 2010-10-20 07:00:00 24.0 897 2010-10-20 -Chicago New York 2010-10-20 07:00:00 15.0 1531 2010-10-20 -Chicago New York 2010-10-20 07:00:00 -6.0 1610 2010-10-20 -Chicago New York 2010-10-20 07:00:00 -2.0 3198 2010-10-20 Cleveland New York 2010-10-20 07:00:00 -8.0 2630 2010-10-20 Cleveland New York 2010-10-20 07:00:00 -15.0 3014 2010-10-20 -Washington New York 2010-10-20 07:00:00 -2.0 7291 2010-10-20 -Baltimore New York 2010-10-22 07:00:00 -12.0 1064 2010-10-22 +Chicago New York 2010-10-20 07:00:00 -2.0 3198 2010-10-20 +Baltimore New York 2010-10-20 07:00:00 -30.0 1064 2010-10-20 +Chicago New York 2010-10-20 07:00:00 -6.0 1610 2010-10-20 +Chicago New York 2010-10-20 07:00:00 15.0 1531 2010-10-20 +Chicago New York 2010-10-22 07:00:00 -11.0 3198 2010-10-22 +Cleveland New York 2010-10-22 07:00:00 -3.0 3014 2010-10-22 Baltimore New York 2010-10-22 07:00:00 54.0 1142 2010-10-22 -Baltimore New York 2010-10-22 07:00:00 18.0 1599 2010-10-22 -Chicago New York 2010-10-22 07:00:00 2.0 361 2010-10-22 -Chicago New York 2010-10-22 07:00:00 24.0 897 2010-10-22 +Baltimore New York 2010-10-22 07:00:00 -12.0 1064 2010-10-22 +Cleveland New York 2010-10-22 07:00:00 -25.0 2646 2010-10-22 +Cleveland New York 2010-10-22 07:00:00 1.0 2630 2010-10-22 Chicago New York 2010-10-22 07:00:00 16.0 1531 2010-10-22 +Chicago New York 2010-10-22 07:00:00 24.0 897 2010-10-22 +Chicago New York 2010-10-22 07:00:00 2.0 361 2010-10-22 +Baltimore New York 2010-10-22 07:00:00 18.0 1599 2010-10-22 Chicago New York 2010-10-22 07:00:00 -6.0 1610 2010-10-22 -Chicago New York 2010-10-22 07:00:00 -11.0 3198 2010-10-22 -Cleveland New York 2010-10-22 07:00:00 1.0 2630 2010-10-22 -Cleveland New York 2010-10-22 07:00:00 -25.0 2646 2010-10-22 -Cleveland New York 2010-10-22 07:00:00 -3.0 3014 2010-10-22 -Baltimore New York 2010-10-31 07:00:00 -1.0 1599 2010-10-31 -Baltimore New York 2010-10-31 07:00:00 -14.0 2571 2010-10-31 Chicago New York 2010-10-31 07:00:00 -25.0 361 2010-10-31 Chicago New York 2010-10-31 07:00:00 -18.0 897 2010-10-31 Chicago New York 2010-10-31 07:00:00 -4.0 1531 2010-10-31 Chicago New York 2010-10-31 07:00:00 -22.0 1610 2010-10-31 Chicago New York 2010-10-31 07:00:00 -15.0 3198 2010-10-31 Washington New York 2010-10-31 07:00:00 -18.0 7282 2010-10-31 +Baltimore New York 2010-10-31 07:00:00 -1.0 1599 2010-10-31 +Baltimore New York 2010-10-31 07:00:00 -14.0 2571 2010-10-31 PREHOOK: query: select * from flights_tiny_orc_partitioned_date sort by fl_num, fl_date limit 25 PREHOOK: type: QUERY PREHOOK: Input: default@flights_tiny_orc_partitioned_date @@ -1015,101 +1015,92 @@ POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-30 POSTHOOK: Input: default@flights_tiny_orc_partitioned_date@fl_date=2010-10-31 #### A masked pattern was here #### Baltimore New York 2010-10-26 07:00:00 -22.0 1064 2010-10-26 -Baltimore New York 2010-10-26 07:00:00 123.0 1142 2010-10-26 -Baltimore New York 2010-10-26 07:00:00 90.0 1599 2010-10-26 -Chicago New York 2010-10-26 07:00:00 12.0 361 2010-10-26 -Chicago New York 2010-10-26 07:00:00 0.0 897 2010-10-26 -Chicago New York 2010-10-26 07:00:00 29.0 1531 2010-10-26 -Chicago New York 2010-10-26 07:00:00 -17.0 1610 2010-10-26 -Chicago New York 2010-10-26 07:00:00 6.0 3198 2010-10-26 +Cleveland New York 2010-10-26 07:00:00 13.0 3014 2010-10-26 Cleveland New York 2010-10-26 07:00:00 4.0 2630 2010-10-26 Cleveland New York 2010-10-26 07:00:00 -27.0 2646 2010-10-26 Cleveland New York 2010-10-26 07:00:00 -11.0 2662 2010-10-26 -Cleveland New York 2010-10-26 07:00:00 13.0 3014 2010-10-26 Washington New York 2010-10-26 07:00:00 4.0 7291 2010-10-26 -Baltimore New York 2010-10-21 07:00:00 17.0 1064 2010-10-21 -Baltimore New York 2010-10-21 07:00:00 105.0 1142 2010-10-21 -Baltimore New York 2010-10-21 07:00:00 28.0 1599 2010-10-21 -Chicago New York 2010-10-21 07:00:00 142.0 361 2010-10-21 -Chicago New York 2010-10-21 07:00:00 77.0 897 2010-10-21 -Chicago New York 2010-10-21 07:00:00 53.0 1531 2010-10-21 -Chicago New York 2010-10-21 07:00:00 -5.0 1610 2010-10-21 -Chicago New York 2010-10-21 07:00:00 51.0 3198 2010-10-21 -Cleveland New York 2010-10-21 07:00:00 3.0 2630 2010-10-21 -Cleveland New York 2010-10-21 07:00:00 29.0 2646 2010-10-21 -Cleveland New York 2010-10-21 07:00:00 72.0 3014 2010-10-21 -Washington New York 2010-10-21 07:00:00 22.0 7291 2010-10-21 +Chicago New York 2010-10-26 07:00:00 6.0 3198 2010-10-26 +Chicago New York 2010-10-26 07:00:00 -17.0 1610 2010-10-26 +Chicago New York 2010-10-26 07:00:00 29.0 1531 2010-10-26 +Chicago New York 2010-10-26 07:00:00 0.0 897 2010-10-26 +Chicago New York 2010-10-26 07:00:00 12.0 361 2010-10-26 +Baltimore New York 2010-10-26 07:00:00 90.0 1599 2010-10-26 +Baltimore New York 2010-10-26 07:00:00 123.0 1142 2010-10-26 +Chicago New York 2010-10-25 07:00:00 6.0 1610 2010-10-25 +Chicago New York 2010-10-25 07:00:00 -16.0 3198 2010-10-25 Baltimore New York 2010-10-25 07:00:00 -25.0 1064 2010-10-25 Baltimore New York 2010-10-25 07:00:00 92.0 1142 2010-10-25 Baltimore New York 2010-10-25 07:00:00 106.0 1599 2010-10-25 Chicago New York 2010-10-25 07:00:00 31.0 361 2010-10-25 -Chicago New York 2010-10-25 07:00:00 -1.0 897 2010-10-25 -Chicago New York 2010-10-25 07:00:00 43.0 1531 2010-10-25 -Chicago New York 2010-10-25 07:00:00 6.0 1610 2010-10-25 -Chicago New York 2010-10-25 07:00:00 -16.0 3198 2010-10-25 +Cleveland New York 2010-10-25 07:00:00 42.0 3014 2010-10-25 Cleveland New York 2010-10-25 07:00:00 -4.0 2630 2010-10-25 Cleveland New York 2010-10-25 07:00:00 81.0 2646 2010-10-25 -Cleveland New York 2010-10-25 07:00:00 42.0 3014 2010-10-25 Washington New York 2010-10-25 07:00:00 9.0 7291 2010-10-25 -Baltimore New York 2010-10-24 07:00:00 12.0 1599 2010-10-24 -Baltimore New York 2010-10-24 07:00:00 20.0 2571 2010-10-24 -Chicago New York 2010-10-24 07:00:00 10.0 361 2010-10-24 -Chicago New York 2010-10-24 07:00:00 113.0 897 2010-10-24 +Chicago New York 2010-10-25 07:00:00 -1.0 897 2010-10-25 +Chicago New York 2010-10-25 07:00:00 43.0 1531 2010-10-25 +Chicago New York 2010-10-21 07:00:00 142.0 361 2010-10-21 +Chicago New York 2010-10-21 07:00:00 77.0 897 2010-10-21 +Chicago New York 2010-10-21 07:00:00 53.0 1531 2010-10-21 +Chicago New York 2010-10-21 07:00:00 -5.0 1610 2010-10-21 +Chicago New York 2010-10-21 07:00:00 51.0 3198 2010-10-21 +Washington New York 2010-10-21 07:00:00 22.0 7291 2010-10-21 +Baltimore New York 2010-10-21 07:00:00 28.0 1599 2010-10-21 +Cleveland New York 2010-10-21 07:00:00 72.0 3014 2010-10-21 +Cleveland New York 2010-10-21 07:00:00 3.0 2630 2010-10-21 +Cleveland New York 2010-10-21 07:00:00 29.0 2646 2010-10-21 +Baltimore New York 2010-10-21 07:00:00 17.0 1064 2010-10-21 +Baltimore New York 2010-10-21 07:00:00 105.0 1142 2010-10-21 Chicago New York 2010-10-24 07:00:00 -5.0 1531 2010-10-24 -Chicago New York 2010-10-24 07:00:00 -17.0 1610 2010-10-24 -Chicago New York 2010-10-24 07:00:00 -3.0 3198 2010-10-24 +Chicago New York 2010-10-24 07:00:00 113.0 897 2010-10-24 +Chicago New York 2010-10-24 07:00:00 10.0 361 2010-10-24 +Baltimore New York 2010-10-24 07:00:00 20.0 2571 2010-10-24 +Baltimore New York 2010-10-24 07:00:00 12.0 1599 2010-10-24 +Washington New York 2010-10-24 07:00:00 -26.0 7282 2010-10-24 Cleveland New York 2010-10-24 07:00:00 5.0 2254 2010-10-24 -Cleveland New York 2010-10-24 07:00:00 -11.0 2630 2010-10-24 -Cleveland New York 2010-10-24 07:00:00 -20.0 2646 2010-10-24 Cleveland New York 2010-10-24 07:00:00 -9.0 3014 2010-10-24 -Washington New York 2010-10-24 07:00:00 -26.0 7282 2010-10-24 -Baltimore New York 2010-10-23 07:00:00 18.0 272 2010-10-23 -Baltimore New York 2010-10-23 07:00:00 -10.0 1805 2010-10-23 -Baltimore New York 2010-10-23 07:00:00 6.0 3171 2010-10-23 -Chicago New York 2010-10-23 07:00:00 3.0 384 2010-10-23 -Chicago New York 2010-10-23 07:00:00 32.0 426 2010-10-23 -Chicago New York 2010-10-23 07:00:00 1.0 650 2010-10-23 -Chicago New York 2010-10-23 07:00:00 11.0 3085 2010-10-23 -Cleveland New York 2010-10-23 07:00:00 -21.0 2932 2010-10-23 -Washington New York 2010-10-23 07:00:00 -25.0 5832 2010-10-23 -Washington New York 2010-10-23 07:00:00 -21.0 5904 2010-10-23 -Washington New York 2010-10-23 07:00:00 -18.0 5917 2010-10-23 -Washington New York 2010-10-23 07:00:00 -16.0 7274 2010-10-23 +Cleveland New York 2010-10-24 07:00:00 -20.0 2646 2010-10-24 +Cleveland New York 2010-10-24 07:00:00 -11.0 2630 2010-10-24 +Chicago New York 2010-10-24 07:00:00 -3.0 3198 2010-10-24 +Chicago New York 2010-10-24 07:00:00 -17.0 1610 2010-10-24 +Chicago New York 2010-10-29 07:00:00 -18.0 1610 2010-10-29 +Cleveland New York 2010-10-29 07:00:00 -12.0 3014 2010-10-29 +Cleveland New York 2010-10-29 07:00:00 -19.0 2646 2010-10-29 +Cleveland New York 2010-10-29 07:00:00 -4.0 2630 2010-10-29 +Washington New York 2010-10-29 07:00:00 1.0 7291 2010-10-29 Baltimore New York 2010-10-29 07:00:00 -24.0 1064 2010-10-29 Baltimore New York 2010-10-29 07:00:00 21.0 1142 2010-10-29 Baltimore New York 2010-10-29 07:00:00 -2.0 1599 2010-10-29 Chicago New York 2010-10-29 07:00:00 -12.0 361 2010-10-29 Chicago New York 2010-10-29 07:00:00 -11.0 897 2010-10-29 Chicago New York 2010-10-29 07:00:00 15.0 1531 2010-10-29 -Chicago New York 2010-10-29 07:00:00 -18.0 1610 2010-10-29 Chicago New York 2010-10-29 07:00:00 -4.0 3198 2010-10-29 -Cleveland New York 2010-10-29 07:00:00 -4.0 2630 2010-10-29 -Cleveland New York 2010-10-29 07:00:00 -19.0 2646 2010-10-29 -Cleveland New York 2010-10-29 07:00:00 -12.0 3014 2010-10-29 -Washington New York 2010-10-29 07:00:00 1.0 7291 2010-10-29 -Baltimore New York 2010-10-27 07:00:00 -18.0 1064 2010-10-27 -Baltimore New York 2010-10-27 07:00:00 49.0 1142 2010-10-27 -Baltimore New York 2010-10-27 07:00:00 92.0 1599 2010-10-27 -Chicago New York 2010-10-27 07:00:00 148.0 361 2010-10-27 -Chicago New York 2010-10-27 07:00:00 -11.0 897 2010-10-27 -Chicago New York 2010-10-27 07:00:00 70.0 1531 2010-10-27 -Chicago New York 2010-10-27 07:00:00 8.0 1610 2010-10-27 -Chicago New York 2010-10-27 07:00:00 21.0 3198 2010-10-27 -Cleveland New York 2010-10-27 07:00:00 16.0 2630 2010-10-27 -Cleveland New York 2010-10-27 07:00:00 27.0 3014 2010-10-27 -Washington New York 2010-10-27 07:00:00 26.0 7291 2010-10-27 +Chicago New York 2010-10-28 07:00:00 3.0 1610 2010-10-28 Baltimore New York 2010-10-28 07:00:00 -4.0 1064 2010-10-28 Baltimore New York 2010-10-28 07:00:00 -14.0 1142 2010-10-28 Baltimore New York 2010-10-28 07:00:00 -14.0 1599 2010-10-28 Chicago New York 2010-10-28 07:00:00 2.0 361 2010-10-28 Chicago New York 2010-10-28 07:00:00 2.0 897 2010-10-28 Chicago New York 2010-10-28 07:00:00 -11.0 1531 2010-10-28 -Chicago New York 2010-10-28 07:00:00 3.0 1610 2010-10-28 Chicago New York 2010-10-28 07:00:00 -18.0 3198 2010-10-28 Cleveland New York 2010-10-28 07:00:00 3.0 2630 2010-10-28 Cleveland New York 2010-10-28 07:00:00 -6.0 2646 2010-10-28 Cleveland New York 2010-10-28 07:00:00 1.0 3014 2010-10-28 Washington New York 2010-10-28 07:00:00 45.0 7291 2010-10-28 +Cleveland New York 2010-10-27 07:00:00 16.0 2630 2010-10-27 +Chicago New York 2010-10-27 07:00:00 21.0 3198 2010-10-27 +Chicago New York 2010-10-27 07:00:00 8.0 1610 2010-10-27 +Chicago New York 2010-10-27 07:00:00 70.0 1531 2010-10-27 +Chicago New York 2010-10-27 07:00:00 -11.0 897 2010-10-27 +Chicago New York 2010-10-27 07:00:00 148.0 361 2010-10-27 +Baltimore New York 2010-10-27 07:00:00 92.0 1599 2010-10-27 +Baltimore New York 2010-10-27 07:00:00 49.0 1142 2010-10-27 +Baltimore New York 2010-10-27 07:00:00 -18.0 1064 2010-10-27 +Washington New York 2010-10-27 07:00:00 26.0 7291 2010-10-27 +Cleveland New York 2010-10-27 07:00:00 27.0 3014 2010-10-27 +Cleveland New York 2010-10-30 07:00:00 -12.0 2932 2010-10-30 +Washington New York 2010-10-30 07:00:00 -16.0 5917 2010-10-30 +Washington New York 2010-10-30 07:00:00 -27.0 5904 2010-10-30 Baltimore New York 2010-10-30 07:00:00 14.0 272 2010-10-30 Baltimore New York 2010-10-30 07:00:00 -1.0 1805 2010-10-30 Baltimore New York 2010-10-30 07:00:00 5.0 3171 2010-10-30 @@ -1118,39 +1109,48 @@ Chicago New York 2010-10-30 07:00:00 -10.0 426 2010-10-30 Chicago New York 2010-10-30 07:00:00 -5.0 650 2010-10-30 Chicago New York 2010-10-30 07:00:00 -5.0 3085 2010-10-30 Cleveland New York 2010-10-30 07:00:00 -23.0 2018 2010-10-30 -Cleveland New York 2010-10-30 07:00:00 -12.0 2932 2010-10-30 -Washington New York 2010-10-30 07:00:00 -27.0 5904 2010-10-30 -Washington New York 2010-10-30 07:00:00 -16.0 5917 2010-10-30 -Baltimore New York 2010-10-20 07:00:00 -30.0 1064 2010-10-20 -Baltimore New York 2010-10-20 07:00:00 23.0 1142 2010-10-20 +Baltimore New York 2010-10-23 07:00:00 6.0 3171 2010-10-23 +Chicago New York 2010-10-23 07:00:00 1.0 650 2010-10-23 +Washington New York 2010-10-23 07:00:00 -25.0 5832 2010-10-23 +Baltimore New York 2010-10-23 07:00:00 -10.0 1805 2010-10-23 +Baltimore New York 2010-10-23 07:00:00 18.0 272 2010-10-23 +Washington New York 2010-10-23 07:00:00 -21.0 5904 2010-10-23 +Washington New York 2010-10-23 07:00:00 -18.0 5917 2010-10-23 +Chicago New York 2010-10-23 07:00:00 32.0 426 2010-10-23 +Cleveland New York 2010-10-23 07:00:00 -21.0 2932 2010-10-23 +Chicago New York 2010-10-23 07:00:00 11.0 3085 2010-10-23 +Washington New York 2010-10-23 07:00:00 -16.0 7274 2010-10-23 +Chicago New York 2010-10-23 07:00:00 3.0 384 2010-10-23 +Chicago New York 2010-10-20 07:00:00 24.0 897 2010-10-20 Baltimore New York 2010-10-20 07:00:00 6.0 1599 2010-10-20 +Baltimore New York 2010-10-20 07:00:00 23.0 1142 2010-10-20 +Washington New York 2010-10-20 07:00:00 -2.0 7291 2010-10-20 Chicago New York 2010-10-20 07:00:00 42.0 361 2010-10-20 -Chicago New York 2010-10-20 07:00:00 24.0 897 2010-10-20 -Chicago New York 2010-10-20 07:00:00 15.0 1531 2010-10-20 -Chicago New York 2010-10-20 07:00:00 -6.0 1610 2010-10-20 -Chicago New York 2010-10-20 07:00:00 -2.0 3198 2010-10-20 Cleveland New York 2010-10-20 07:00:00 -8.0 2630 2010-10-20 Cleveland New York 2010-10-20 07:00:00 -15.0 3014 2010-10-20 -Washington New York 2010-10-20 07:00:00 -2.0 7291 2010-10-20 -Baltimore New York 2010-10-22 07:00:00 -12.0 1064 2010-10-22 +Chicago New York 2010-10-20 07:00:00 -2.0 3198 2010-10-20 +Baltimore New York 2010-10-20 07:00:00 -30.0 1064 2010-10-20 +Chicago New York 2010-10-20 07:00:00 -6.0 1610 2010-10-20 +Chicago New York 2010-10-20 07:00:00 15.0 1531 2010-10-20 +Chicago New York 2010-10-22 07:00:00 -11.0 3198 2010-10-22 +Cleveland New York 2010-10-22 07:00:00 -3.0 3014 2010-10-22 Baltimore New York 2010-10-22 07:00:00 54.0 1142 2010-10-22 -Baltimore New York 2010-10-22 07:00:00 18.0 1599 2010-10-22 -Chicago New York 2010-10-22 07:00:00 2.0 361 2010-10-22 -Chicago New York 2010-10-22 07:00:00 24.0 897 2010-10-22 +Baltimore New York 2010-10-22 07:00:00 -12.0 1064 2010-10-22 +Cleveland New York 2010-10-22 07:00:00 -25.0 2646 2010-10-22 +Cleveland New York 2010-10-22 07:00:00 1.0 2630 2010-10-22 Chicago New York 2010-10-22 07:00:00 16.0 1531 2010-10-22 +Chicago New York 2010-10-22 07:00:00 24.0 897 2010-10-22 +Chicago New York 2010-10-22 07:00:00 2.0 361 2010-10-22 +Baltimore New York 2010-10-22 07:00:00 18.0 1599 2010-10-22 Chicago New York 2010-10-22 07:00:00 -6.0 1610 2010-10-22 -Chicago New York 2010-10-22 07:00:00 -11.0 3198 2010-10-22 -Cleveland New York 2010-10-22 07:00:00 1.0 2630 2010-10-22 -Cleveland New York 2010-10-22 07:00:00 -25.0 2646 2010-10-22 -Cleveland New York 2010-10-22 07:00:00 -3.0 3014 2010-10-22 -Baltimore New York 2010-10-31 07:00:00 -1.0 1599 2010-10-31 -Baltimore New York 2010-10-31 07:00:00 -14.0 2571 2010-10-31 Chicago New York 2010-10-31 07:00:00 -25.0 361 2010-10-31 Chicago New York 2010-10-31 07:00:00 -18.0 897 2010-10-31 Chicago New York 2010-10-31 07:00:00 -4.0 1531 2010-10-31 Chicago New York 2010-10-31 07:00:00 -22.0 1610 2010-10-31 Chicago New York 2010-10-31 07:00:00 -15.0 3198 2010-10-31 Washington New York 2010-10-31 07:00:00 -18.0 7282 2010-10-31 +Baltimore New York 2010-10-31 07:00:00 -1.0 1599 2010-10-31 +Baltimore New York 2010-10-31 07:00:00 -14.0 2571 2010-10-31 PREHOOK: query: explain vectorization expression select * from flights_tiny_orc_partitioned_date sort by fl_num, fl_date limit 25 PREHOOK: type: QUERY @@ -1637,6 +1637,20 @@ POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10- POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-30 07%3A00%3A00 POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-31 07%3A00%3A00 #### A masked pattern was here #### +Chicago New York 2010-10-25 -16.0 3198 2010-10-25 07:00:00 +Cleveland New York 2010-10-25 81.0 2646 2010-10-25 07:00:00 +Cleveland New York 2010-10-25 -4.0 2630 2010-10-25 07:00:00 +Washington New York 2010-10-25 9.0 7291 2010-10-25 07:00:00 +Cleveland New York 2010-10-25 42.0 3014 2010-10-25 07:00:00 +Baltimore New York 2010-10-25 -25.0 1064 2010-10-25 07:00:00 +Baltimore New York 2010-10-25 92.0 1142 2010-10-25 07:00:00 +Baltimore New York 2010-10-25 106.0 1599 2010-10-25 07:00:00 +Chicago New York 2010-10-25 31.0 361 2010-10-25 07:00:00 +Chicago New York 2010-10-25 -1.0 897 2010-10-25 07:00:00 +Chicago New York 2010-10-25 43.0 1531 2010-10-25 07:00:00 +Chicago New York 2010-10-25 6.0 1610 2010-10-25 07:00:00 +Cleveland New York 2010-10-26 4.0 2630 2010-10-26 07:00:00 +Washington New York 2010-10-26 4.0 7291 2010-10-26 07:00:00 Baltimore New York 2010-10-26 -22.0 1064 2010-10-26 07:00:00 Baltimore New York 2010-10-26 123.0 1142 2010-10-26 07:00:00 Baltimore New York 2010-10-26 90.0 1599 2010-10-26 07:00:00 @@ -1645,95 +1659,80 @@ Chicago New York 2010-10-26 0.0 897 2010-10-26 07:00:00 Chicago New York 2010-10-26 29.0 1531 2010-10-26 07:00:00 Chicago New York 2010-10-26 -17.0 1610 2010-10-26 07:00:00 Chicago New York 2010-10-26 6.0 3198 2010-10-26 07:00:00 -Cleveland New York 2010-10-26 4.0 2630 2010-10-26 07:00:00 -Cleveland New York 2010-10-26 -27.0 2646 2010-10-26 07:00:00 -Cleveland New York 2010-10-26 -11.0 2662 2010-10-26 07:00:00 Cleveland New York 2010-10-26 13.0 3014 2010-10-26 07:00:00 -Washington New York 2010-10-26 4.0 7291 2010-10-26 07:00:00 -Baltimore New York 2010-10-21 17.0 1064 2010-10-21 07:00:00 -Baltimore New York 2010-10-21 105.0 1142 2010-10-21 07:00:00 -Baltimore New York 2010-10-21 28.0 1599 2010-10-21 07:00:00 -Chicago New York 2010-10-21 142.0 361 2010-10-21 07:00:00 -Chicago New York 2010-10-21 77.0 897 2010-10-21 07:00:00 -Chicago New York 2010-10-21 53.0 1531 2010-10-21 07:00:00 +Cleveland New York 2010-10-26 -11.0 2662 2010-10-26 07:00:00 +Cleveland New York 2010-10-26 -27.0 2646 2010-10-26 07:00:00 Chicago New York 2010-10-21 -5.0 1610 2010-10-21 07:00:00 Chicago New York 2010-10-21 51.0 3198 2010-10-21 07:00:00 -Cleveland New York 2010-10-21 3.0 2630 2010-10-21 07:00:00 -Cleveland New York 2010-10-21 29.0 2646 2010-10-21 07:00:00 Cleveland New York 2010-10-21 72.0 3014 2010-10-21 07:00:00 +Cleveland New York 2010-10-21 29.0 2646 2010-10-21 07:00:00 +Baltimore New York 2010-10-21 17.0 1064 2010-10-21 07:00:00 +Cleveland New York 2010-10-21 3.0 2630 2010-10-21 07:00:00 +Baltimore New York 2010-10-21 28.0 1599 2010-10-21 07:00:00 +Chicago New York 2010-10-21 77.0 897 2010-10-21 07:00:00 Washington New York 2010-10-21 22.0 7291 2010-10-21 07:00:00 -Baltimore New York 2010-10-25 -25.0 1064 2010-10-25 07:00:00 -Baltimore New York 2010-10-25 92.0 1142 2010-10-25 07:00:00 -Baltimore New York 2010-10-25 106.0 1599 2010-10-25 07:00:00 -Chicago New York 2010-10-25 31.0 361 2010-10-25 07:00:00 -Chicago New York 2010-10-25 -1.0 897 2010-10-25 07:00:00 -Chicago New York 2010-10-25 43.0 1531 2010-10-25 07:00:00 -Chicago New York 2010-10-25 6.0 1610 2010-10-25 07:00:00 -Chicago New York 2010-10-25 -16.0 3198 2010-10-25 07:00:00 -Cleveland New York 2010-10-25 -4.0 2630 2010-10-25 07:00:00 -Cleveland New York 2010-10-25 81.0 2646 2010-10-25 07:00:00 -Cleveland New York 2010-10-25 42.0 3014 2010-10-25 07:00:00 -Washington New York 2010-10-25 9.0 7291 2010-10-25 07:00:00 +Chicago New York 2010-10-21 142.0 361 2010-10-21 07:00:00 +Chicago New York 2010-10-21 53.0 1531 2010-10-21 07:00:00 +Baltimore New York 2010-10-21 105.0 1142 2010-10-21 07:00:00 +Washington New York 2010-10-23 -16.0 7274 2010-10-23 07:00:00 +Baltimore New York 2010-10-23 18.0 272 2010-10-23 07:00:00 +Baltimore New York 2010-10-23 -10.0 1805 2010-10-23 07:00:00 +Baltimore New York 2010-10-23 6.0 3171 2010-10-23 07:00:00 +Washington New York 2010-10-23 -18.0 5917 2010-10-23 07:00:00 +Washington New York 2010-10-23 -21.0 5904 2010-10-23 07:00:00 +Washington New York 2010-10-23 -25.0 5832 2010-10-23 07:00:00 +Cleveland New York 2010-10-23 -21.0 2932 2010-10-23 07:00:00 +Chicago New York 2010-10-23 11.0 3085 2010-10-23 07:00:00 +Chicago New York 2010-10-23 1.0 650 2010-10-23 07:00:00 +Chicago New York 2010-10-23 32.0 426 2010-10-23 07:00:00 +Chicago New York 2010-10-23 3.0 384 2010-10-23 07:00:00 +Chicago New York 2010-10-24 -17.0 1610 2010-10-24 07:00:00 +Washington New York 2010-10-24 -26.0 7282 2010-10-24 07:00:00 Baltimore New York 2010-10-24 12.0 1599 2010-10-24 07:00:00 Baltimore New York 2010-10-24 20.0 2571 2010-10-24 07:00:00 -Chicago New York 2010-10-24 10.0 361 2010-10-24 07:00:00 Chicago New York 2010-10-24 113.0 897 2010-10-24 07:00:00 -Chicago New York 2010-10-24 -5.0 1531 2010-10-24 07:00:00 -Chicago New York 2010-10-24 -17.0 1610 2010-10-24 07:00:00 Chicago New York 2010-10-24 -3.0 3198 2010-10-24 07:00:00 Cleveland New York 2010-10-24 5.0 2254 2010-10-24 07:00:00 Cleveland New York 2010-10-24 -11.0 2630 2010-10-24 07:00:00 Cleveland New York 2010-10-24 -20.0 2646 2010-10-24 07:00:00 -Cleveland New York 2010-10-24 -9.0 3014 2010-10-24 07:00:00 -Washington New York 2010-10-24 -26.0 7282 2010-10-24 07:00:00 -Baltimore New York 2010-10-23 18.0 272 2010-10-23 07:00:00 -Baltimore New York 2010-10-23 -10.0 1805 2010-10-23 07:00:00 -Baltimore New York 2010-10-23 6.0 3171 2010-10-23 07:00:00 -Chicago New York 2010-10-23 3.0 384 2010-10-23 07:00:00 -Chicago New York 2010-10-23 32.0 426 2010-10-23 07:00:00 -Chicago New York 2010-10-23 1.0 650 2010-10-23 07:00:00 -Chicago New York 2010-10-23 11.0 3085 2010-10-23 07:00:00 -Cleveland New York 2010-10-23 -21.0 2932 2010-10-23 07:00:00 -Washington New York 2010-10-23 -25.0 5832 2010-10-23 07:00:00 -Washington New York 2010-10-23 -21.0 5904 2010-10-23 07:00:00 -Washington New York 2010-10-23 -18.0 5917 2010-10-23 07:00:00 -Washington New York 2010-10-23 -16.0 7274 2010-10-23 07:00:00 -Baltimore New York 2010-10-29 -24.0 1064 2010-10-29 07:00:00 -Baltimore New York 2010-10-29 21.0 1142 2010-10-29 07:00:00 -Baltimore New York 2010-10-29 -2.0 1599 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -12.0 361 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -11.0 897 2010-10-29 07:00:00 -Chicago New York 2010-10-29 15.0 1531 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -18.0 1610 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -4.0 3198 2010-10-29 07:00:00 -Cleveland New York 2010-10-29 -4.0 2630 2010-10-29 07:00:00 -Cleveland New York 2010-10-29 -19.0 2646 2010-10-29 07:00:00 -Cleveland New York 2010-10-29 -12.0 3014 2010-10-29 07:00:00 +Cleveland New York 2010-10-24 -9.0 3014 2010-10-24 07:00:00 +Chicago New York 2010-10-24 10.0 361 2010-10-24 07:00:00 +Chicago New York 2010-10-24 -5.0 1531 2010-10-24 07:00:00 +Baltimore New York 2010-10-29 21.0 1142 2010-10-29 07:00:00 Washington New York 2010-10-29 1.0 7291 2010-10-29 07:00:00 -Baltimore New York 2010-10-27 -18.0 1064 2010-10-27 07:00:00 -Baltimore New York 2010-10-27 49.0 1142 2010-10-27 07:00:00 -Baltimore New York 2010-10-27 92.0 1599 2010-10-27 07:00:00 -Chicago New York 2010-10-27 148.0 361 2010-10-27 07:00:00 -Chicago New York 2010-10-27 -11.0 897 2010-10-27 07:00:00 -Chicago New York 2010-10-27 70.0 1531 2010-10-27 07:00:00 -Chicago New York 2010-10-27 8.0 1610 2010-10-27 07:00:00 +Cleveland New York 2010-10-29 -12.0 3014 2010-10-29 07:00:00 +Cleveland New York 2010-10-29 -19.0 2646 2010-10-29 07:00:00 +Cleveland New York 2010-10-29 -4.0 2630 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -4.0 3198 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -18.0 1610 2010-10-29 07:00:00 +Chicago New York 2010-10-29 15.0 1531 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -11.0 897 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -12.0 361 2010-10-29 07:00:00 +Baltimore New York 2010-10-29 -2.0 1599 2010-10-29 07:00:00 +Baltimore New York 2010-10-29 -24.0 1064 2010-10-29 07:00:00 +Washington New York 2010-10-27 26.0 7291 2010-10-27 07:00:00 Chicago New York 2010-10-27 21.0 3198 2010-10-27 07:00:00 +Chicago New York 2010-10-27 8.0 1610 2010-10-27 07:00:00 Cleveland New York 2010-10-27 16.0 2630 2010-10-27 07:00:00 Cleveland New York 2010-10-27 27.0 3014 2010-10-27 07:00:00 -Washington New York 2010-10-27 26.0 7291 2010-10-27 07:00:00 +Chicago New York 2010-10-27 -11.0 897 2010-10-27 07:00:00 +Chicago New York 2010-10-27 148.0 361 2010-10-27 07:00:00 +Baltimore New York 2010-10-27 92.0 1599 2010-10-27 07:00:00 +Baltimore New York 2010-10-27 49.0 1142 2010-10-27 07:00:00 +Baltimore New York 2010-10-27 -18.0 1064 2010-10-27 07:00:00 +Chicago New York 2010-10-27 70.0 1531 2010-10-27 07:00:00 +Cleveland New York 2010-10-28 -6.0 2646 2010-10-28 07:00:00 +Cleveland New York 2010-10-28 3.0 2630 2010-10-28 07:00:00 Baltimore New York 2010-10-28 -4.0 1064 2010-10-28 07:00:00 +Chicago New York 2010-10-28 3.0 1610 2010-10-28 07:00:00 +Washington New York 2010-10-28 45.0 7291 2010-10-28 07:00:00 Baltimore New York 2010-10-28 -14.0 1142 2010-10-28 07:00:00 Baltimore New York 2010-10-28 -14.0 1599 2010-10-28 07:00:00 Chicago New York 2010-10-28 2.0 361 2010-10-28 07:00:00 Chicago New York 2010-10-28 2.0 897 2010-10-28 07:00:00 Chicago New York 2010-10-28 -11.0 1531 2010-10-28 07:00:00 -Chicago New York 2010-10-28 3.0 1610 2010-10-28 07:00:00 Chicago New York 2010-10-28 -18.0 3198 2010-10-28 07:00:00 -Cleveland New York 2010-10-28 3.0 2630 2010-10-28 07:00:00 -Cleveland New York 2010-10-28 -6.0 2646 2010-10-28 07:00:00 Cleveland New York 2010-10-28 1.0 3014 2010-10-28 07:00:00 -Washington New York 2010-10-28 45.0 7291 2010-10-28 07:00:00 -Baltimore New York 2010-10-30 14.0 272 2010-10-30 07:00:00 Baltimore New York 2010-10-30 -1.0 1805 2010-10-30 07:00:00 Baltimore New York 2010-10-30 5.0 3171 2010-10-30 07:00:00 Chicago New York 2010-10-30 -6.0 384 2010-10-30 07:00:00 @@ -1744,28 +1743,30 @@ Cleveland New York 2010-10-30 -23.0 2018 2010-10-30 07:00:00 Cleveland New York 2010-10-30 -12.0 2932 2010-10-30 07:00:00 Washington New York 2010-10-30 -27.0 5904 2010-10-30 07:00:00 Washington New York 2010-10-30 -16.0 5917 2010-10-30 07:00:00 -Baltimore New York 2010-10-20 -30.0 1064 2010-10-20 07:00:00 +Baltimore New York 2010-10-30 14.0 272 2010-10-30 07:00:00 +Chicago New York 2010-10-20 42.0 361 2010-10-20 07:00:00 +Washington New York 2010-10-20 -2.0 7291 2010-10-20 07:00:00 +Chicago New York 2010-10-20 -2.0 3198 2010-10-20 07:00:00 Baltimore New York 2010-10-20 23.0 1142 2010-10-20 07:00:00 Baltimore New York 2010-10-20 6.0 1599 2010-10-20 07:00:00 -Chicago New York 2010-10-20 42.0 361 2010-10-20 07:00:00 +Baltimore New York 2010-10-20 -30.0 1064 2010-10-20 07:00:00 +Cleveland New York 2010-10-20 -8.0 2630 2010-10-20 07:00:00 +Cleveland New York 2010-10-20 -15.0 3014 2010-10-20 07:00:00 Chicago New York 2010-10-20 24.0 897 2010-10-20 07:00:00 Chicago New York 2010-10-20 15.0 1531 2010-10-20 07:00:00 Chicago New York 2010-10-20 -6.0 1610 2010-10-20 07:00:00 -Chicago New York 2010-10-20 -2.0 3198 2010-10-20 07:00:00 -Cleveland New York 2010-10-20 -8.0 2630 2010-10-20 07:00:00 -Cleveland New York 2010-10-20 -15.0 3014 2010-10-20 07:00:00 -Washington New York 2010-10-20 -2.0 7291 2010-10-20 07:00:00 +Cleveland New York 2010-10-22 1.0 2630 2010-10-22 07:00:00 Baltimore New York 2010-10-22 -12.0 1064 2010-10-22 07:00:00 -Baltimore New York 2010-10-22 54.0 1142 2010-10-22 07:00:00 -Baltimore New York 2010-10-22 18.0 1599 2010-10-22 07:00:00 -Chicago New York 2010-10-22 2.0 361 2010-10-22 07:00:00 -Chicago New York 2010-10-22 24.0 897 2010-10-22 07:00:00 -Chicago New York 2010-10-22 16.0 1531 2010-10-22 07:00:00 -Chicago New York 2010-10-22 -6.0 1610 2010-10-22 07:00:00 Chicago New York 2010-10-22 -11.0 3198 2010-10-22 07:00:00 -Cleveland New York 2010-10-22 1.0 2630 2010-10-22 07:00:00 -Cleveland New York 2010-10-22 -25.0 2646 2010-10-22 07:00:00 +Chicago New York 2010-10-22 -6.0 1610 2010-10-22 07:00:00 +Chicago New York 2010-10-22 16.0 1531 2010-10-22 07:00:00 +Chicago New York 2010-10-22 24.0 897 2010-10-22 07:00:00 +Chicago New York 2010-10-22 2.0 361 2010-10-22 07:00:00 +Baltimore New York 2010-10-22 18.0 1599 2010-10-22 07:00:00 +Baltimore New York 2010-10-22 54.0 1142 2010-10-22 07:00:00 Cleveland New York 2010-10-22 -3.0 3014 2010-10-22 07:00:00 +Cleveland New York 2010-10-22 -25.0 2646 2010-10-22 07:00:00 +Washington New York 2010-10-31 -18.0 7282 2010-10-31 07:00:00 Baltimore New York 2010-10-31 -1.0 1599 2010-10-31 07:00:00 Baltimore New York 2010-10-31 -14.0 2571 2010-10-31 07:00:00 Chicago New York 2010-10-31 -25.0 361 2010-10-31 07:00:00 @@ -1773,7 +1774,6 @@ Chicago New York 2010-10-31 -18.0 897 2010-10-31 07:00:00 Chicago New York 2010-10-31 -4.0 1531 2010-10-31 07:00:00 Chicago New York 2010-10-31 -22.0 1610 2010-10-31 07:00:00 Chicago New York 2010-10-31 -15.0 3198 2010-10-31 07:00:00 -Washington New York 2010-10-31 -18.0 7282 2010-10-31 07:00:00 PREHOOK: query: select * from flights_tiny_orc_partitioned_timestamp sort by fl_num, fl_time limit 25 PREHOOK: type: QUERY PREHOOK: Input: default@flights_tiny_orc_partitioned_timestamp @@ -1969,6 +1969,20 @@ POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10- POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-30 07%3A00%3A00 POSTHOOK: Input: default@flights_tiny_orc_partitioned_timestamp@fl_time=2010-10-31 07%3A00%3A00 #### A masked pattern was here #### +Chicago New York 2010-10-25 -16.0 3198 2010-10-25 07:00:00 +Cleveland New York 2010-10-25 81.0 2646 2010-10-25 07:00:00 +Cleveland New York 2010-10-25 -4.0 2630 2010-10-25 07:00:00 +Washington New York 2010-10-25 9.0 7291 2010-10-25 07:00:00 +Cleveland New York 2010-10-25 42.0 3014 2010-10-25 07:00:00 +Baltimore New York 2010-10-25 -25.0 1064 2010-10-25 07:00:00 +Baltimore New York 2010-10-25 92.0 1142 2010-10-25 07:00:00 +Baltimore New York 2010-10-25 106.0 1599 2010-10-25 07:00:00 +Chicago New York 2010-10-25 31.0 361 2010-10-25 07:00:00 +Chicago New York 2010-10-25 -1.0 897 2010-10-25 07:00:00 +Chicago New York 2010-10-25 43.0 1531 2010-10-25 07:00:00 +Chicago New York 2010-10-25 6.0 1610 2010-10-25 07:00:00 +Cleveland New York 2010-10-26 4.0 2630 2010-10-26 07:00:00 +Washington New York 2010-10-26 4.0 7291 2010-10-26 07:00:00 Baltimore New York 2010-10-26 -22.0 1064 2010-10-26 07:00:00 Baltimore New York 2010-10-26 123.0 1142 2010-10-26 07:00:00 Baltimore New York 2010-10-26 90.0 1599 2010-10-26 07:00:00 @@ -1977,95 +1991,80 @@ Chicago New York 2010-10-26 0.0 897 2010-10-26 07:00:00 Chicago New York 2010-10-26 29.0 1531 2010-10-26 07:00:00 Chicago New York 2010-10-26 -17.0 1610 2010-10-26 07:00:00 Chicago New York 2010-10-26 6.0 3198 2010-10-26 07:00:00 -Cleveland New York 2010-10-26 4.0 2630 2010-10-26 07:00:00 -Cleveland New York 2010-10-26 -27.0 2646 2010-10-26 07:00:00 -Cleveland New York 2010-10-26 -11.0 2662 2010-10-26 07:00:00 Cleveland New York 2010-10-26 13.0 3014 2010-10-26 07:00:00 -Washington New York 2010-10-26 4.0 7291 2010-10-26 07:00:00 -Baltimore New York 2010-10-21 17.0 1064 2010-10-21 07:00:00 -Baltimore New York 2010-10-21 105.0 1142 2010-10-21 07:00:00 -Baltimore New York 2010-10-21 28.0 1599 2010-10-21 07:00:00 -Chicago New York 2010-10-21 142.0 361 2010-10-21 07:00:00 -Chicago New York 2010-10-21 77.0 897 2010-10-21 07:00:00 -Chicago New York 2010-10-21 53.0 1531 2010-10-21 07:00:00 +Cleveland New York 2010-10-26 -11.0 2662 2010-10-26 07:00:00 +Cleveland New York 2010-10-26 -27.0 2646 2010-10-26 07:00:00 Chicago New York 2010-10-21 -5.0 1610 2010-10-21 07:00:00 Chicago New York 2010-10-21 51.0 3198 2010-10-21 07:00:00 -Cleveland New York 2010-10-21 3.0 2630 2010-10-21 07:00:00 -Cleveland New York 2010-10-21 29.0 2646 2010-10-21 07:00:00 Cleveland New York 2010-10-21 72.0 3014 2010-10-21 07:00:00 +Cleveland New York 2010-10-21 29.0 2646 2010-10-21 07:00:00 +Baltimore New York 2010-10-21 17.0 1064 2010-10-21 07:00:00 +Cleveland New York 2010-10-21 3.0 2630 2010-10-21 07:00:00 +Baltimore New York 2010-10-21 28.0 1599 2010-10-21 07:00:00 +Chicago New York 2010-10-21 77.0 897 2010-10-21 07:00:00 Washington New York 2010-10-21 22.0 7291 2010-10-21 07:00:00 -Baltimore New York 2010-10-25 -25.0 1064 2010-10-25 07:00:00 -Baltimore New York 2010-10-25 92.0 1142 2010-10-25 07:00:00 -Baltimore New York 2010-10-25 106.0 1599 2010-10-25 07:00:00 -Chicago New York 2010-10-25 31.0 361 2010-10-25 07:00:00 -Chicago New York 2010-10-25 -1.0 897 2010-10-25 07:00:00 -Chicago New York 2010-10-25 43.0 1531 2010-10-25 07:00:00 -Chicago New York 2010-10-25 6.0 1610 2010-10-25 07:00:00 -Chicago New York 2010-10-25 -16.0 3198 2010-10-25 07:00:00 -Cleveland New York 2010-10-25 -4.0 2630 2010-10-25 07:00:00 -Cleveland New York 2010-10-25 81.0 2646 2010-10-25 07:00:00 -Cleveland New York 2010-10-25 42.0 3014 2010-10-25 07:00:00 -Washington New York 2010-10-25 9.0 7291 2010-10-25 07:00:00 +Chicago New York 2010-10-21 142.0 361 2010-10-21 07:00:00 +Chicago New York 2010-10-21 53.0 1531 2010-10-21 07:00:00 +Baltimore New York 2010-10-21 105.0 1142 2010-10-21 07:00:00 +Washington New York 2010-10-23 -16.0 7274 2010-10-23 07:00:00 +Baltimore New York 2010-10-23 18.0 272 2010-10-23 07:00:00 +Baltimore New York 2010-10-23 -10.0 1805 2010-10-23 07:00:00 +Baltimore New York 2010-10-23 6.0 3171 2010-10-23 07:00:00 +Washington New York 2010-10-23 -18.0 5917 2010-10-23 07:00:00 +Washington New York 2010-10-23 -21.0 5904 2010-10-23 07:00:00 +Washington New York 2010-10-23 -25.0 5832 2010-10-23 07:00:00 +Cleveland New York 2010-10-23 -21.0 2932 2010-10-23 07:00:00 +Chicago New York 2010-10-23 11.0 3085 2010-10-23 07:00:00 +Chicago New York 2010-10-23 1.0 650 2010-10-23 07:00:00 +Chicago New York 2010-10-23 32.0 426 2010-10-23 07:00:00 +Chicago New York 2010-10-23 3.0 384 2010-10-23 07:00:00 +Chicago New York 2010-10-24 -17.0 1610 2010-10-24 07:00:00 +Washington New York 2010-10-24 -26.0 7282 2010-10-24 07:00:00 Baltimore New York 2010-10-24 12.0 1599 2010-10-24 07:00:00 Baltimore New York 2010-10-24 20.0 2571 2010-10-24 07:00:00 -Chicago New York 2010-10-24 10.0 361 2010-10-24 07:00:00 Chicago New York 2010-10-24 113.0 897 2010-10-24 07:00:00 -Chicago New York 2010-10-24 -5.0 1531 2010-10-24 07:00:00 -Chicago New York 2010-10-24 -17.0 1610 2010-10-24 07:00:00 Chicago New York 2010-10-24 -3.0 3198 2010-10-24 07:00:00 Cleveland New York 2010-10-24 5.0 2254 2010-10-24 07:00:00 Cleveland New York 2010-10-24 -11.0 2630 2010-10-24 07:00:00 Cleveland New York 2010-10-24 -20.0 2646 2010-10-24 07:00:00 Cleveland New York 2010-10-24 -9.0 3014 2010-10-24 07:00:00 -Washington New York 2010-10-24 -26.0 7282 2010-10-24 07:00:00 -Baltimore New York 2010-10-23 18.0 272 2010-10-23 07:00:00 -Baltimore New York 2010-10-23 -10.0 1805 2010-10-23 07:00:00 -Baltimore New York 2010-10-23 6.0 3171 2010-10-23 07:00:00 -Chicago New York 2010-10-23 3.0 384 2010-10-23 07:00:00 -Chicago New York 2010-10-23 32.0 426 2010-10-23 07:00:00 -Chicago New York 2010-10-23 1.0 650 2010-10-23 07:00:00 -Chicago New York 2010-10-23 11.0 3085 2010-10-23 07:00:00 -Cleveland New York 2010-10-23 -21.0 2932 2010-10-23 07:00:00 -Washington New York 2010-10-23 -25.0 5832 2010-10-23 07:00:00 -Washington New York 2010-10-23 -21.0 5904 2010-10-23 07:00:00 -Washington New York 2010-10-23 -18.0 5917 2010-10-23 07:00:00 -Washington New York 2010-10-23 -16.0 7274 2010-10-23 07:00:00 -Baltimore New York 2010-10-29 -24.0 1064 2010-10-29 07:00:00 +Chicago New York 2010-10-24 10.0 361 2010-10-24 07:00:00 +Chicago New York 2010-10-24 -5.0 1531 2010-10-24 07:00:00 Baltimore New York 2010-10-29 21.0 1142 2010-10-29 07:00:00 -Baltimore New York 2010-10-29 -2.0 1599 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -12.0 361 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -11.0 897 2010-10-29 07:00:00 -Chicago New York 2010-10-29 15.0 1531 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -18.0 1610 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -4.0 3198 2010-10-29 07:00:00 -Cleveland New York 2010-10-29 -4.0 2630 2010-10-29 07:00:00 -Cleveland New York 2010-10-29 -19.0 2646 2010-10-29 07:00:00 -Cleveland New York 2010-10-29 -12.0 3014 2010-10-29 07:00:00 Washington New York 2010-10-29 1.0 7291 2010-10-29 07:00:00 -Baltimore New York 2010-10-27 -18.0 1064 2010-10-27 07:00:00 -Baltimore New York 2010-10-27 49.0 1142 2010-10-27 07:00:00 -Baltimore New York 2010-10-27 92.0 1599 2010-10-27 07:00:00 -Chicago New York 2010-10-27 148.0 361 2010-10-27 07:00:00 -Chicago New York 2010-10-27 -11.0 897 2010-10-27 07:00:00 -Chicago New York 2010-10-27 70.0 1531 2010-10-27 07:00:00 -Chicago New York 2010-10-27 8.0 1610 2010-10-27 07:00:00 +Cleveland New York 2010-10-29 -12.0 3014 2010-10-29 07:00:00 +Cleveland New York 2010-10-29 -19.0 2646 2010-10-29 07:00:00 +Cleveland New York 2010-10-29 -4.0 2630 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -4.0 3198 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -18.0 1610 2010-10-29 07:00:00 +Chicago New York 2010-10-29 15.0 1531 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -11.0 897 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -12.0 361 2010-10-29 07:00:00 +Baltimore New York 2010-10-29 -2.0 1599 2010-10-29 07:00:00 +Baltimore New York 2010-10-29 -24.0 1064 2010-10-29 07:00:00 +Washington New York 2010-10-27 26.0 7291 2010-10-27 07:00:00 Chicago New York 2010-10-27 21.0 3198 2010-10-27 07:00:00 +Chicago New York 2010-10-27 8.0 1610 2010-10-27 07:00:00 Cleveland New York 2010-10-27 16.0 2630 2010-10-27 07:00:00 Cleveland New York 2010-10-27 27.0 3014 2010-10-27 07:00:00 -Washington New York 2010-10-27 26.0 7291 2010-10-27 07:00:00 +Chicago New York 2010-10-27 -11.0 897 2010-10-27 07:00:00 +Chicago New York 2010-10-27 148.0 361 2010-10-27 07:00:00 +Baltimore New York 2010-10-27 92.0 1599 2010-10-27 07:00:00 +Baltimore New York 2010-10-27 49.0 1142 2010-10-27 07:00:00 +Baltimore New York 2010-10-27 -18.0 1064 2010-10-27 07:00:00 +Chicago New York 2010-10-27 70.0 1531 2010-10-27 07:00:00 +Cleveland New York 2010-10-28 -6.0 2646 2010-10-28 07:00:00 +Cleveland New York 2010-10-28 3.0 2630 2010-10-28 07:00:00 Baltimore New York 2010-10-28 -4.0 1064 2010-10-28 07:00:00 +Chicago New York 2010-10-28 3.0 1610 2010-10-28 07:00:00 +Washington New York 2010-10-28 45.0 7291 2010-10-28 07:00:00 Baltimore New York 2010-10-28 -14.0 1142 2010-10-28 07:00:00 Baltimore New York 2010-10-28 -14.0 1599 2010-10-28 07:00:00 Chicago New York 2010-10-28 2.0 361 2010-10-28 07:00:00 Chicago New York 2010-10-28 2.0 897 2010-10-28 07:00:00 Chicago New York 2010-10-28 -11.0 1531 2010-10-28 07:00:00 -Chicago New York 2010-10-28 3.0 1610 2010-10-28 07:00:00 Chicago New York 2010-10-28 -18.0 3198 2010-10-28 07:00:00 -Cleveland New York 2010-10-28 3.0 2630 2010-10-28 07:00:00 -Cleveland New York 2010-10-28 -6.0 2646 2010-10-28 07:00:00 Cleveland New York 2010-10-28 1.0 3014 2010-10-28 07:00:00 -Washington New York 2010-10-28 45.0 7291 2010-10-28 07:00:00 -Baltimore New York 2010-10-30 14.0 272 2010-10-30 07:00:00 Baltimore New York 2010-10-30 -1.0 1805 2010-10-30 07:00:00 Baltimore New York 2010-10-30 5.0 3171 2010-10-30 07:00:00 Chicago New York 2010-10-30 -6.0 384 2010-10-30 07:00:00 @@ -2076,28 +2075,30 @@ Cleveland New York 2010-10-30 -23.0 2018 2010-10-30 07:00:00 Cleveland New York 2010-10-30 -12.0 2932 2010-10-30 07:00:00 Washington New York 2010-10-30 -27.0 5904 2010-10-30 07:00:00 Washington New York 2010-10-30 -16.0 5917 2010-10-30 07:00:00 -Baltimore New York 2010-10-20 -30.0 1064 2010-10-20 07:00:00 +Baltimore New York 2010-10-30 14.0 272 2010-10-30 07:00:00 +Chicago New York 2010-10-20 42.0 361 2010-10-20 07:00:00 +Washington New York 2010-10-20 -2.0 7291 2010-10-20 07:00:00 +Chicago New York 2010-10-20 -2.0 3198 2010-10-20 07:00:00 Baltimore New York 2010-10-20 23.0 1142 2010-10-20 07:00:00 Baltimore New York 2010-10-20 6.0 1599 2010-10-20 07:00:00 -Chicago New York 2010-10-20 42.0 361 2010-10-20 07:00:00 +Baltimore New York 2010-10-20 -30.0 1064 2010-10-20 07:00:00 +Cleveland New York 2010-10-20 -8.0 2630 2010-10-20 07:00:00 +Cleveland New York 2010-10-20 -15.0 3014 2010-10-20 07:00:00 Chicago New York 2010-10-20 24.0 897 2010-10-20 07:00:00 Chicago New York 2010-10-20 15.0 1531 2010-10-20 07:00:00 Chicago New York 2010-10-20 -6.0 1610 2010-10-20 07:00:00 -Chicago New York 2010-10-20 -2.0 3198 2010-10-20 07:00:00 -Cleveland New York 2010-10-20 -8.0 2630 2010-10-20 07:00:00 -Cleveland New York 2010-10-20 -15.0 3014 2010-10-20 07:00:00 -Washington New York 2010-10-20 -2.0 7291 2010-10-20 07:00:00 +Cleveland New York 2010-10-22 1.0 2630 2010-10-22 07:00:00 Baltimore New York 2010-10-22 -12.0 1064 2010-10-22 07:00:00 -Baltimore New York 2010-10-22 54.0 1142 2010-10-22 07:00:00 -Baltimore New York 2010-10-22 18.0 1599 2010-10-22 07:00:00 -Chicago New York 2010-10-22 2.0 361 2010-10-22 07:00:00 -Chicago New York 2010-10-22 24.0 897 2010-10-22 07:00:00 -Chicago New York 2010-10-22 16.0 1531 2010-10-22 07:00:00 -Chicago New York 2010-10-22 -6.0 1610 2010-10-22 07:00:00 Chicago New York 2010-10-22 -11.0 3198 2010-10-22 07:00:00 -Cleveland New York 2010-10-22 1.0 2630 2010-10-22 07:00:00 -Cleveland New York 2010-10-22 -25.0 2646 2010-10-22 07:00:00 +Chicago New York 2010-10-22 -6.0 1610 2010-10-22 07:00:00 +Chicago New York 2010-10-22 16.0 1531 2010-10-22 07:00:00 +Chicago New York 2010-10-22 24.0 897 2010-10-22 07:00:00 +Chicago New York 2010-10-22 2.0 361 2010-10-22 07:00:00 +Baltimore New York 2010-10-22 18.0 1599 2010-10-22 07:00:00 +Baltimore New York 2010-10-22 54.0 1142 2010-10-22 07:00:00 Cleveland New York 2010-10-22 -3.0 3014 2010-10-22 07:00:00 +Cleveland New York 2010-10-22 -25.0 2646 2010-10-22 07:00:00 +Washington New York 2010-10-31 -18.0 7282 2010-10-31 07:00:00 Baltimore New York 2010-10-31 -1.0 1599 2010-10-31 07:00:00 Baltimore New York 2010-10-31 -14.0 2571 2010-10-31 07:00:00 Chicago New York 2010-10-31 -25.0 361 2010-10-31 07:00:00 @@ -2105,7 +2106,6 @@ Chicago New York 2010-10-31 -18.0 897 2010-10-31 07:00:00 Chicago New York 2010-10-31 -4.0 1531 2010-10-31 07:00:00 Chicago New York 2010-10-31 -22.0 1610 2010-10-31 07:00:00 Chicago New York 2010-10-31 -15.0 3198 2010-10-31 07:00:00 -Washington New York 2010-10-31 -18.0 7282 2010-10-31 07:00:00 PREHOOK: query: explain vectorization expression select * from flights_tiny_orc_partitioned_timestamp sort by fl_num, fl_time limit 25 PREHOOK: type: QUERY @@ -3107,113 +3107,116 @@ POSTHOOK: Input: default@flights_tiny_parquet_partitioned_date@fl_date=2010-10-2 POSTHOOK: Input: default@flights_tiny_parquet_partitioned_date@fl_date=2010-10-30 POSTHOOK: Input: default@flights_tiny_parquet_partitioned_date@fl_date=2010-10-31 #### A masked pattern was here #### +Chicago New York 2010-10-28 07:00:00 3.0 1610 2010-10-28 Baltimore New York 2010-10-28 07:00:00 -4.0 1064 2010-10-28 Baltimore New York 2010-10-28 07:00:00 -14.0 1142 2010-10-28 Baltimore New York 2010-10-28 07:00:00 -14.0 1599 2010-10-28 Chicago New York 2010-10-28 07:00:00 2.0 361 2010-10-28 Chicago New York 2010-10-28 07:00:00 2.0 897 2010-10-28 Chicago New York 2010-10-28 07:00:00 -11.0 1531 2010-10-28 -Chicago New York 2010-10-28 07:00:00 3.0 1610 2010-10-28 Chicago New York 2010-10-28 07:00:00 -18.0 3198 2010-10-28 Cleveland New York 2010-10-28 07:00:00 3.0 2630 2010-10-28 Cleveland New York 2010-10-28 07:00:00 -6.0 2646 2010-10-28 Cleveland New York 2010-10-28 07:00:00 1.0 3014 2010-10-28 Washington New York 2010-10-28 07:00:00 45.0 7291 2010-10-28 Baltimore New York 2010-10-26 07:00:00 -22.0 1064 2010-10-26 -Baltimore New York 2010-10-26 07:00:00 123.0 1142 2010-10-26 -Baltimore New York 2010-10-26 07:00:00 90.0 1599 2010-10-26 -Chicago New York 2010-10-26 07:00:00 12.0 361 2010-10-26 -Chicago New York 2010-10-26 07:00:00 0.0 897 2010-10-26 -Chicago New York 2010-10-26 07:00:00 29.0 1531 2010-10-26 -Chicago New York 2010-10-26 07:00:00 -17.0 1610 2010-10-26 -Chicago New York 2010-10-26 07:00:00 6.0 3198 2010-10-26 +Cleveland New York 2010-10-26 07:00:00 13.0 3014 2010-10-26 Cleveland New York 2010-10-26 07:00:00 4.0 2630 2010-10-26 Cleveland New York 2010-10-26 07:00:00 -27.0 2646 2010-10-26 Cleveland New York 2010-10-26 07:00:00 -11.0 2662 2010-10-26 -Cleveland New York 2010-10-26 07:00:00 13.0 3014 2010-10-26 Washington New York 2010-10-26 07:00:00 4.0 7291 2010-10-26 -Baltimore New York 2010-10-21 07:00:00 17.0 1064 2010-10-21 -Baltimore New York 2010-10-21 07:00:00 105.0 1142 2010-10-21 -Baltimore New York 2010-10-21 07:00:00 28.0 1599 2010-10-21 +Chicago New York 2010-10-26 07:00:00 6.0 3198 2010-10-26 +Chicago New York 2010-10-26 07:00:00 -17.0 1610 2010-10-26 +Chicago New York 2010-10-26 07:00:00 29.0 1531 2010-10-26 +Chicago New York 2010-10-26 07:00:00 0.0 897 2010-10-26 +Chicago New York 2010-10-26 07:00:00 12.0 361 2010-10-26 +Baltimore New York 2010-10-26 07:00:00 90.0 1599 2010-10-26 +Baltimore New York 2010-10-26 07:00:00 123.0 1142 2010-10-26 Chicago New York 2010-10-21 07:00:00 142.0 361 2010-10-21 Chicago New York 2010-10-21 07:00:00 77.0 897 2010-10-21 Chicago New York 2010-10-21 07:00:00 53.0 1531 2010-10-21 Chicago New York 2010-10-21 07:00:00 -5.0 1610 2010-10-21 Chicago New York 2010-10-21 07:00:00 51.0 3198 2010-10-21 +Washington New York 2010-10-21 07:00:00 22.0 7291 2010-10-21 +Baltimore New York 2010-10-21 07:00:00 28.0 1599 2010-10-21 +Cleveland New York 2010-10-21 07:00:00 72.0 3014 2010-10-21 Cleveland New York 2010-10-21 07:00:00 3.0 2630 2010-10-21 Cleveland New York 2010-10-21 07:00:00 29.0 2646 2010-10-21 -Cleveland New York 2010-10-21 07:00:00 72.0 3014 2010-10-21 -Washington New York 2010-10-21 07:00:00 22.0 7291 2010-10-21 -Baltimore New York 2010-10-23 07:00:00 18.0 272 2010-10-23 -Baltimore New York 2010-10-23 07:00:00 -10.0 1805 2010-10-23 +Baltimore New York 2010-10-21 07:00:00 17.0 1064 2010-10-21 +Baltimore New York 2010-10-21 07:00:00 105.0 1142 2010-10-21 Baltimore New York 2010-10-23 07:00:00 6.0 3171 2010-10-23 -Chicago New York 2010-10-23 07:00:00 3.0 384 2010-10-23 -Chicago New York 2010-10-23 07:00:00 32.0 426 2010-10-23 Chicago New York 2010-10-23 07:00:00 1.0 650 2010-10-23 -Chicago New York 2010-10-23 07:00:00 11.0 3085 2010-10-23 -Cleveland New York 2010-10-23 07:00:00 -21.0 2932 2010-10-23 Washington New York 2010-10-23 07:00:00 -25.0 5832 2010-10-23 +Baltimore New York 2010-10-23 07:00:00 -10.0 1805 2010-10-23 +Baltimore New York 2010-10-23 07:00:00 18.0 272 2010-10-23 Washington New York 2010-10-23 07:00:00 -21.0 5904 2010-10-23 Washington New York 2010-10-23 07:00:00 -18.0 5917 2010-10-23 +Chicago New York 2010-10-23 07:00:00 32.0 426 2010-10-23 +Cleveland New York 2010-10-23 07:00:00 -21.0 2932 2010-10-23 +Chicago New York 2010-10-23 07:00:00 11.0 3085 2010-10-23 Washington New York 2010-10-23 07:00:00 -16.0 7274 2010-10-23 -Baltimore New York 2010-10-24 07:00:00 12.0 1599 2010-10-24 -Baltimore New York 2010-10-24 07:00:00 20.0 2571 2010-10-24 -Chicago New York 2010-10-24 07:00:00 10.0 361 2010-10-24 -Chicago New York 2010-10-24 07:00:00 113.0 897 2010-10-24 +Chicago New York 2010-10-23 07:00:00 3.0 384 2010-10-23 Chicago New York 2010-10-24 07:00:00 -5.0 1531 2010-10-24 -Chicago New York 2010-10-24 07:00:00 -17.0 1610 2010-10-24 -Chicago New York 2010-10-24 07:00:00 -3.0 3198 2010-10-24 +Chicago New York 2010-10-24 07:00:00 113.0 897 2010-10-24 +Chicago New York 2010-10-24 07:00:00 10.0 361 2010-10-24 +Baltimore New York 2010-10-24 07:00:00 20.0 2571 2010-10-24 +Baltimore New York 2010-10-24 07:00:00 12.0 1599 2010-10-24 +Washington New York 2010-10-24 07:00:00 -26.0 7282 2010-10-24 Cleveland New York 2010-10-24 07:00:00 5.0 2254 2010-10-24 -Cleveland New York 2010-10-24 07:00:00 -11.0 2630 2010-10-24 -Cleveland New York 2010-10-24 07:00:00 -20.0 2646 2010-10-24 Cleveland New York 2010-10-24 07:00:00 -9.0 3014 2010-10-24 -Washington New York 2010-10-24 07:00:00 -26.0 7282 2010-10-24 +Cleveland New York 2010-10-24 07:00:00 -20.0 2646 2010-10-24 +Cleveland New York 2010-10-24 07:00:00 -11.0 2630 2010-10-24 +Chicago New York 2010-10-24 07:00:00 -3.0 3198 2010-10-24 +Chicago New York 2010-10-24 07:00:00 -17.0 1610 2010-10-24 +Chicago New York 2010-10-25 07:00:00 6.0 1610 2010-10-25 +Chicago New York 2010-10-25 07:00:00 -16.0 3198 2010-10-25 Baltimore New York 2010-10-25 07:00:00 -25.0 1064 2010-10-25 Baltimore New York 2010-10-25 07:00:00 92.0 1142 2010-10-25 Baltimore New York 2010-10-25 07:00:00 106.0 1599 2010-10-25 Chicago New York 2010-10-25 07:00:00 31.0 361 2010-10-25 -Chicago New York 2010-10-25 07:00:00 -1.0 897 2010-10-25 -Chicago New York 2010-10-25 07:00:00 43.0 1531 2010-10-25 -Chicago New York 2010-10-25 07:00:00 6.0 1610 2010-10-25 -Chicago New York 2010-10-25 07:00:00 -16.0 3198 2010-10-25 +Cleveland New York 2010-10-25 07:00:00 42.0 3014 2010-10-25 Cleveland New York 2010-10-25 07:00:00 -4.0 2630 2010-10-25 Cleveland New York 2010-10-25 07:00:00 81.0 2646 2010-10-25 -Cleveland New York 2010-10-25 07:00:00 42.0 3014 2010-10-25 Washington New York 2010-10-25 07:00:00 9.0 7291 2010-10-25 +Chicago New York 2010-10-25 07:00:00 -1.0 897 2010-10-25 +Chicago New York 2010-10-25 07:00:00 43.0 1531 2010-10-25 +Chicago New York 2010-10-29 07:00:00 -18.0 1610 2010-10-29 +Cleveland New York 2010-10-29 07:00:00 -12.0 3014 2010-10-29 +Cleveland New York 2010-10-29 07:00:00 -19.0 2646 2010-10-29 +Cleveland New York 2010-10-29 07:00:00 -4.0 2630 2010-10-29 +Washington New York 2010-10-29 07:00:00 1.0 7291 2010-10-29 Baltimore New York 2010-10-29 07:00:00 -24.0 1064 2010-10-29 Baltimore New York 2010-10-29 07:00:00 21.0 1142 2010-10-29 Baltimore New York 2010-10-29 07:00:00 -2.0 1599 2010-10-29 Chicago New York 2010-10-29 07:00:00 -12.0 361 2010-10-29 Chicago New York 2010-10-29 07:00:00 -11.0 897 2010-10-29 Chicago New York 2010-10-29 07:00:00 15.0 1531 2010-10-29 -Chicago New York 2010-10-29 07:00:00 -18.0 1610 2010-10-29 Chicago New York 2010-10-29 07:00:00 -4.0 3198 2010-10-29 -Cleveland New York 2010-10-29 07:00:00 -4.0 2630 2010-10-29 -Cleveland New York 2010-10-29 07:00:00 -19.0 2646 2010-10-29 -Cleveland New York 2010-10-29 07:00:00 -12.0 3014 2010-10-29 -Washington New York 2010-10-29 07:00:00 1.0 7291 2010-10-29 -Baltimore New York 2010-10-20 07:00:00 -30.0 1064 2010-10-20 -Baltimore New York 2010-10-20 07:00:00 23.0 1142 2010-10-20 +Chicago New York 2010-10-20 07:00:00 24.0 897 2010-10-20 Baltimore New York 2010-10-20 07:00:00 6.0 1599 2010-10-20 +Baltimore New York 2010-10-20 07:00:00 23.0 1142 2010-10-20 +Washington New York 2010-10-20 07:00:00 -2.0 7291 2010-10-20 Chicago New York 2010-10-20 07:00:00 42.0 361 2010-10-20 -Chicago New York 2010-10-20 07:00:00 24.0 897 2010-10-20 -Chicago New York 2010-10-20 07:00:00 15.0 1531 2010-10-20 -Chicago New York 2010-10-20 07:00:00 -6.0 1610 2010-10-20 -Chicago New York 2010-10-20 07:00:00 -2.0 3198 2010-10-20 Cleveland New York 2010-10-20 07:00:00 -8.0 2630 2010-10-20 Cleveland New York 2010-10-20 07:00:00 -15.0 3014 2010-10-20 -Washington New York 2010-10-20 07:00:00 -2.0 7291 2010-10-20 -Baltimore New York 2010-10-27 07:00:00 -18.0 1064 2010-10-27 -Baltimore New York 2010-10-27 07:00:00 49.0 1142 2010-10-27 -Baltimore New York 2010-10-27 07:00:00 92.0 1599 2010-10-27 -Chicago New York 2010-10-27 07:00:00 148.0 361 2010-10-27 -Chicago New York 2010-10-27 07:00:00 -11.0 897 2010-10-27 -Chicago New York 2010-10-27 07:00:00 70.0 1531 2010-10-27 -Chicago New York 2010-10-27 07:00:00 8.0 1610 2010-10-27 -Chicago New York 2010-10-27 07:00:00 21.0 3198 2010-10-27 +Chicago New York 2010-10-20 07:00:00 -2.0 3198 2010-10-20 +Baltimore New York 2010-10-20 07:00:00 -30.0 1064 2010-10-20 +Chicago New York 2010-10-20 07:00:00 -6.0 1610 2010-10-20 +Chicago New York 2010-10-20 07:00:00 15.0 1531 2010-10-20 Cleveland New York 2010-10-27 07:00:00 16.0 2630 2010-10-27 -Cleveland New York 2010-10-27 07:00:00 27.0 3014 2010-10-27 +Chicago New York 2010-10-27 07:00:00 21.0 3198 2010-10-27 +Chicago New York 2010-10-27 07:00:00 8.0 1610 2010-10-27 +Chicago New York 2010-10-27 07:00:00 70.0 1531 2010-10-27 +Chicago New York 2010-10-27 07:00:00 -11.0 897 2010-10-27 +Chicago New York 2010-10-27 07:00:00 148.0 361 2010-10-27 +Baltimore New York 2010-10-27 07:00:00 92.0 1599 2010-10-27 +Baltimore New York 2010-10-27 07:00:00 49.0 1142 2010-10-27 +Baltimore New York 2010-10-27 07:00:00 -18.0 1064 2010-10-27 Washington New York 2010-10-27 07:00:00 26.0 7291 2010-10-27 +Cleveland New York 2010-10-27 07:00:00 27.0 3014 2010-10-27 +Cleveland New York 2010-10-30 07:00:00 -12.0 2932 2010-10-30 +Washington New York 2010-10-30 07:00:00 -16.0 5917 2010-10-30 +Washington New York 2010-10-30 07:00:00 -27.0 5904 2010-10-30 Baltimore New York 2010-10-30 07:00:00 14.0 272 2010-10-30 Baltimore New York 2010-10-30 07:00:00 -1.0 1805 2010-10-30 Baltimore New York 2010-10-30 07:00:00 5.0 3171 2010-10-30 @@ -3222,28 +3225,25 @@ Chicago New York 2010-10-30 07:00:00 -10.0 426 2010-10-30 Chicago New York 2010-10-30 07:00:00 -5.0 650 2010-10-30 Chicago New York 2010-10-30 07:00:00 -5.0 3085 2010-10-30 Cleveland New York 2010-10-30 07:00:00 -23.0 2018 2010-10-30 -Cleveland New York 2010-10-30 07:00:00 -12.0 2932 2010-10-30 -Washington New York 2010-10-30 07:00:00 -27.0 5904 2010-10-30 -Washington New York 2010-10-30 07:00:00 -16.0 5917 2010-10-30 -Baltimore New York 2010-10-22 07:00:00 -12.0 1064 2010-10-22 +Chicago New York 2010-10-22 07:00:00 -11.0 3198 2010-10-22 +Cleveland New York 2010-10-22 07:00:00 -3.0 3014 2010-10-22 Baltimore New York 2010-10-22 07:00:00 54.0 1142 2010-10-22 -Baltimore New York 2010-10-22 07:00:00 18.0 1599 2010-10-22 -Chicago New York 2010-10-22 07:00:00 2.0 361 2010-10-22 -Chicago New York 2010-10-22 07:00:00 24.0 897 2010-10-22 +Baltimore New York 2010-10-22 07:00:00 -12.0 1064 2010-10-22 +Cleveland New York 2010-10-22 07:00:00 -25.0 2646 2010-10-22 +Cleveland New York 2010-10-22 07:00:00 1.0 2630 2010-10-22 Chicago New York 2010-10-22 07:00:00 16.0 1531 2010-10-22 +Chicago New York 2010-10-22 07:00:00 24.0 897 2010-10-22 +Chicago New York 2010-10-22 07:00:00 2.0 361 2010-10-22 +Baltimore New York 2010-10-22 07:00:00 18.0 1599 2010-10-22 Chicago New York 2010-10-22 07:00:00 -6.0 1610 2010-10-22 -Chicago New York 2010-10-22 07:00:00 -11.0 3198 2010-10-22 -Cleveland New York 2010-10-22 07:00:00 1.0 2630 2010-10-22 -Cleveland New York 2010-10-22 07:00:00 -25.0 2646 2010-10-22 -Cleveland New York 2010-10-22 07:00:00 -3.0 3014 2010-10-22 -Baltimore New York 2010-10-31 07:00:00 -1.0 1599 2010-10-31 -Baltimore New York 2010-10-31 07:00:00 -14.0 2571 2010-10-31 Chicago New York 2010-10-31 07:00:00 -25.0 361 2010-10-31 Chicago New York 2010-10-31 07:00:00 -18.0 897 2010-10-31 Chicago New York 2010-10-31 07:00:00 -4.0 1531 2010-10-31 Chicago New York 2010-10-31 07:00:00 -22.0 1610 2010-10-31 Chicago New York 2010-10-31 07:00:00 -15.0 3198 2010-10-31 Washington New York 2010-10-31 07:00:00 -18.0 7282 2010-10-31 +Baltimore New York 2010-10-31 07:00:00 -1.0 1599 2010-10-31 +Baltimore New York 2010-10-31 07:00:00 -14.0 2571 2010-10-31 PREHOOK: query: select * from flights_tiny_parquet_partitioned_date sort by fl_num, fl_date limit 25 PREHOOK: type: QUERY PREHOOK: Input: default@flights_tiny_parquet_partitioned_date @@ -3439,113 +3439,116 @@ POSTHOOK: Input: default@flights_tiny_parquet_partitioned_date@fl_date=2010-10-2 POSTHOOK: Input: default@flights_tiny_parquet_partitioned_date@fl_date=2010-10-30 POSTHOOK: Input: default@flights_tiny_parquet_partitioned_date@fl_date=2010-10-31 #### A masked pattern was here #### +Chicago New York 2010-10-28 07:00:00 3.0 1610 2010-10-28 Baltimore New York 2010-10-28 07:00:00 -4.0 1064 2010-10-28 Baltimore New York 2010-10-28 07:00:00 -14.0 1142 2010-10-28 Baltimore New York 2010-10-28 07:00:00 -14.0 1599 2010-10-28 Chicago New York 2010-10-28 07:00:00 2.0 361 2010-10-28 Chicago New York 2010-10-28 07:00:00 2.0 897 2010-10-28 Chicago New York 2010-10-28 07:00:00 -11.0 1531 2010-10-28 -Chicago New York 2010-10-28 07:00:00 3.0 1610 2010-10-28 Chicago New York 2010-10-28 07:00:00 -18.0 3198 2010-10-28 Cleveland New York 2010-10-28 07:00:00 3.0 2630 2010-10-28 Cleveland New York 2010-10-28 07:00:00 -6.0 2646 2010-10-28 Cleveland New York 2010-10-28 07:00:00 1.0 3014 2010-10-28 Washington New York 2010-10-28 07:00:00 45.0 7291 2010-10-28 Baltimore New York 2010-10-26 07:00:00 -22.0 1064 2010-10-26 -Baltimore New York 2010-10-26 07:00:00 123.0 1142 2010-10-26 -Baltimore New York 2010-10-26 07:00:00 90.0 1599 2010-10-26 -Chicago New York 2010-10-26 07:00:00 12.0 361 2010-10-26 -Chicago New York 2010-10-26 07:00:00 0.0 897 2010-10-26 -Chicago New York 2010-10-26 07:00:00 29.0 1531 2010-10-26 -Chicago New York 2010-10-26 07:00:00 -17.0 1610 2010-10-26 -Chicago New York 2010-10-26 07:00:00 6.0 3198 2010-10-26 +Cleveland New York 2010-10-26 07:00:00 13.0 3014 2010-10-26 Cleveland New York 2010-10-26 07:00:00 4.0 2630 2010-10-26 Cleveland New York 2010-10-26 07:00:00 -27.0 2646 2010-10-26 Cleveland New York 2010-10-26 07:00:00 -11.0 2662 2010-10-26 -Cleveland New York 2010-10-26 07:00:00 13.0 3014 2010-10-26 Washington New York 2010-10-26 07:00:00 4.0 7291 2010-10-26 -Baltimore New York 2010-10-21 07:00:00 17.0 1064 2010-10-21 -Baltimore New York 2010-10-21 07:00:00 105.0 1142 2010-10-21 -Baltimore New York 2010-10-21 07:00:00 28.0 1599 2010-10-21 +Chicago New York 2010-10-26 07:00:00 6.0 3198 2010-10-26 +Chicago New York 2010-10-26 07:00:00 -17.0 1610 2010-10-26 +Chicago New York 2010-10-26 07:00:00 29.0 1531 2010-10-26 +Chicago New York 2010-10-26 07:00:00 0.0 897 2010-10-26 +Chicago New York 2010-10-26 07:00:00 12.0 361 2010-10-26 +Baltimore New York 2010-10-26 07:00:00 90.0 1599 2010-10-26 +Baltimore New York 2010-10-26 07:00:00 123.0 1142 2010-10-26 Chicago New York 2010-10-21 07:00:00 142.0 361 2010-10-21 Chicago New York 2010-10-21 07:00:00 77.0 897 2010-10-21 Chicago New York 2010-10-21 07:00:00 53.0 1531 2010-10-21 Chicago New York 2010-10-21 07:00:00 -5.0 1610 2010-10-21 Chicago New York 2010-10-21 07:00:00 51.0 3198 2010-10-21 +Washington New York 2010-10-21 07:00:00 22.0 7291 2010-10-21 +Baltimore New York 2010-10-21 07:00:00 28.0 1599 2010-10-21 +Cleveland New York 2010-10-21 07:00:00 72.0 3014 2010-10-21 Cleveland New York 2010-10-21 07:00:00 3.0 2630 2010-10-21 Cleveland New York 2010-10-21 07:00:00 29.0 2646 2010-10-21 -Cleveland New York 2010-10-21 07:00:00 72.0 3014 2010-10-21 -Washington New York 2010-10-21 07:00:00 22.0 7291 2010-10-21 -Baltimore New York 2010-10-23 07:00:00 18.0 272 2010-10-23 -Baltimore New York 2010-10-23 07:00:00 -10.0 1805 2010-10-23 +Baltimore New York 2010-10-21 07:00:00 17.0 1064 2010-10-21 +Baltimore New York 2010-10-21 07:00:00 105.0 1142 2010-10-21 Baltimore New York 2010-10-23 07:00:00 6.0 3171 2010-10-23 -Chicago New York 2010-10-23 07:00:00 3.0 384 2010-10-23 -Chicago New York 2010-10-23 07:00:00 32.0 426 2010-10-23 Chicago New York 2010-10-23 07:00:00 1.0 650 2010-10-23 -Chicago New York 2010-10-23 07:00:00 11.0 3085 2010-10-23 -Cleveland New York 2010-10-23 07:00:00 -21.0 2932 2010-10-23 Washington New York 2010-10-23 07:00:00 -25.0 5832 2010-10-23 +Baltimore New York 2010-10-23 07:00:00 -10.0 1805 2010-10-23 +Baltimore New York 2010-10-23 07:00:00 18.0 272 2010-10-23 Washington New York 2010-10-23 07:00:00 -21.0 5904 2010-10-23 Washington New York 2010-10-23 07:00:00 -18.0 5917 2010-10-23 +Chicago New York 2010-10-23 07:00:00 32.0 426 2010-10-23 +Cleveland New York 2010-10-23 07:00:00 -21.0 2932 2010-10-23 +Chicago New York 2010-10-23 07:00:00 11.0 3085 2010-10-23 Washington New York 2010-10-23 07:00:00 -16.0 7274 2010-10-23 -Baltimore New York 2010-10-24 07:00:00 12.0 1599 2010-10-24 -Baltimore New York 2010-10-24 07:00:00 20.0 2571 2010-10-24 -Chicago New York 2010-10-24 07:00:00 10.0 361 2010-10-24 -Chicago New York 2010-10-24 07:00:00 113.0 897 2010-10-24 +Chicago New York 2010-10-23 07:00:00 3.0 384 2010-10-23 Chicago New York 2010-10-24 07:00:00 -5.0 1531 2010-10-24 -Chicago New York 2010-10-24 07:00:00 -17.0 1610 2010-10-24 -Chicago New York 2010-10-24 07:00:00 -3.0 3198 2010-10-24 +Chicago New York 2010-10-24 07:00:00 113.0 897 2010-10-24 +Chicago New York 2010-10-24 07:00:00 10.0 361 2010-10-24 +Baltimore New York 2010-10-24 07:00:00 20.0 2571 2010-10-24 +Baltimore New York 2010-10-24 07:00:00 12.0 1599 2010-10-24 +Washington New York 2010-10-24 07:00:00 -26.0 7282 2010-10-24 Cleveland New York 2010-10-24 07:00:00 5.0 2254 2010-10-24 -Cleveland New York 2010-10-24 07:00:00 -11.0 2630 2010-10-24 -Cleveland New York 2010-10-24 07:00:00 -20.0 2646 2010-10-24 Cleveland New York 2010-10-24 07:00:00 -9.0 3014 2010-10-24 -Washington New York 2010-10-24 07:00:00 -26.0 7282 2010-10-24 +Cleveland New York 2010-10-24 07:00:00 -20.0 2646 2010-10-24 +Cleveland New York 2010-10-24 07:00:00 -11.0 2630 2010-10-24 +Chicago New York 2010-10-24 07:00:00 -3.0 3198 2010-10-24 +Chicago New York 2010-10-24 07:00:00 -17.0 1610 2010-10-24 +Chicago New York 2010-10-25 07:00:00 6.0 1610 2010-10-25 +Chicago New York 2010-10-25 07:00:00 -16.0 3198 2010-10-25 Baltimore New York 2010-10-25 07:00:00 -25.0 1064 2010-10-25 Baltimore New York 2010-10-25 07:00:00 92.0 1142 2010-10-25 Baltimore New York 2010-10-25 07:00:00 106.0 1599 2010-10-25 Chicago New York 2010-10-25 07:00:00 31.0 361 2010-10-25 -Chicago New York 2010-10-25 07:00:00 -1.0 897 2010-10-25 -Chicago New York 2010-10-25 07:00:00 43.0 1531 2010-10-25 -Chicago New York 2010-10-25 07:00:00 6.0 1610 2010-10-25 -Chicago New York 2010-10-25 07:00:00 -16.0 3198 2010-10-25 +Cleveland New York 2010-10-25 07:00:00 42.0 3014 2010-10-25 Cleveland New York 2010-10-25 07:00:00 -4.0 2630 2010-10-25 Cleveland New York 2010-10-25 07:00:00 81.0 2646 2010-10-25 -Cleveland New York 2010-10-25 07:00:00 42.0 3014 2010-10-25 Washington New York 2010-10-25 07:00:00 9.0 7291 2010-10-25 +Chicago New York 2010-10-25 07:00:00 -1.0 897 2010-10-25 +Chicago New York 2010-10-25 07:00:00 43.0 1531 2010-10-25 +Chicago New York 2010-10-29 07:00:00 -18.0 1610 2010-10-29 +Cleveland New York 2010-10-29 07:00:00 -12.0 3014 2010-10-29 +Cleveland New York 2010-10-29 07:00:00 -19.0 2646 2010-10-29 +Cleveland New York 2010-10-29 07:00:00 -4.0 2630 2010-10-29 +Washington New York 2010-10-29 07:00:00 1.0 7291 2010-10-29 Baltimore New York 2010-10-29 07:00:00 -24.0 1064 2010-10-29 Baltimore New York 2010-10-29 07:00:00 21.0 1142 2010-10-29 Baltimore New York 2010-10-29 07:00:00 -2.0 1599 2010-10-29 Chicago New York 2010-10-29 07:00:00 -12.0 361 2010-10-29 Chicago New York 2010-10-29 07:00:00 -11.0 897 2010-10-29 Chicago New York 2010-10-29 07:00:00 15.0 1531 2010-10-29 -Chicago New York 2010-10-29 07:00:00 -18.0 1610 2010-10-29 Chicago New York 2010-10-29 07:00:00 -4.0 3198 2010-10-29 -Cleveland New York 2010-10-29 07:00:00 -4.0 2630 2010-10-29 -Cleveland New York 2010-10-29 07:00:00 -19.0 2646 2010-10-29 -Cleveland New York 2010-10-29 07:00:00 -12.0 3014 2010-10-29 -Washington New York 2010-10-29 07:00:00 1.0 7291 2010-10-29 -Baltimore New York 2010-10-20 07:00:00 -30.0 1064 2010-10-20 -Baltimore New York 2010-10-20 07:00:00 23.0 1142 2010-10-20 +Chicago New York 2010-10-20 07:00:00 24.0 897 2010-10-20 Baltimore New York 2010-10-20 07:00:00 6.0 1599 2010-10-20 +Baltimore New York 2010-10-20 07:00:00 23.0 1142 2010-10-20 +Washington New York 2010-10-20 07:00:00 -2.0 7291 2010-10-20 Chicago New York 2010-10-20 07:00:00 42.0 361 2010-10-20 -Chicago New York 2010-10-20 07:00:00 24.0 897 2010-10-20 -Chicago New York 2010-10-20 07:00:00 15.0 1531 2010-10-20 -Chicago New York 2010-10-20 07:00:00 -6.0 1610 2010-10-20 -Chicago New York 2010-10-20 07:00:00 -2.0 3198 2010-10-20 Cleveland New York 2010-10-20 07:00:00 -8.0 2630 2010-10-20 Cleveland New York 2010-10-20 07:00:00 -15.0 3014 2010-10-20 -Washington New York 2010-10-20 07:00:00 -2.0 7291 2010-10-20 -Baltimore New York 2010-10-27 07:00:00 -18.0 1064 2010-10-27 -Baltimore New York 2010-10-27 07:00:00 49.0 1142 2010-10-27 -Baltimore New York 2010-10-27 07:00:00 92.0 1599 2010-10-27 -Chicago New York 2010-10-27 07:00:00 148.0 361 2010-10-27 -Chicago New York 2010-10-27 07:00:00 -11.0 897 2010-10-27 -Chicago New York 2010-10-27 07:00:00 70.0 1531 2010-10-27 -Chicago New York 2010-10-27 07:00:00 8.0 1610 2010-10-27 -Chicago New York 2010-10-27 07:00:00 21.0 3198 2010-10-27 +Chicago New York 2010-10-20 07:00:00 -2.0 3198 2010-10-20 +Baltimore New York 2010-10-20 07:00:00 -30.0 1064 2010-10-20 +Chicago New York 2010-10-20 07:00:00 -6.0 1610 2010-10-20 +Chicago New York 2010-10-20 07:00:00 15.0 1531 2010-10-20 Cleveland New York 2010-10-27 07:00:00 16.0 2630 2010-10-27 -Cleveland New York 2010-10-27 07:00:00 27.0 3014 2010-10-27 +Chicago New York 2010-10-27 07:00:00 21.0 3198 2010-10-27 +Chicago New York 2010-10-27 07:00:00 8.0 1610 2010-10-27 +Chicago New York 2010-10-27 07:00:00 70.0 1531 2010-10-27 +Chicago New York 2010-10-27 07:00:00 -11.0 897 2010-10-27 +Chicago New York 2010-10-27 07:00:00 148.0 361 2010-10-27 +Baltimore New York 2010-10-27 07:00:00 92.0 1599 2010-10-27 +Baltimore New York 2010-10-27 07:00:00 49.0 1142 2010-10-27 +Baltimore New York 2010-10-27 07:00:00 -18.0 1064 2010-10-27 Washington New York 2010-10-27 07:00:00 26.0 7291 2010-10-27 +Cleveland New York 2010-10-27 07:00:00 27.0 3014 2010-10-27 +Cleveland New York 2010-10-30 07:00:00 -12.0 2932 2010-10-30 +Washington New York 2010-10-30 07:00:00 -16.0 5917 2010-10-30 +Washington New York 2010-10-30 07:00:00 -27.0 5904 2010-10-30 Baltimore New York 2010-10-30 07:00:00 14.0 272 2010-10-30 Baltimore New York 2010-10-30 07:00:00 -1.0 1805 2010-10-30 Baltimore New York 2010-10-30 07:00:00 5.0 3171 2010-10-30 @@ -3554,28 +3557,25 @@ Chicago New York 2010-10-30 07:00:00 -10.0 426 2010-10-30 Chicago New York 2010-10-30 07:00:00 -5.0 650 2010-10-30 Chicago New York 2010-10-30 07:00:00 -5.0 3085 2010-10-30 Cleveland New York 2010-10-30 07:00:00 -23.0 2018 2010-10-30 -Cleveland New York 2010-10-30 07:00:00 -12.0 2932 2010-10-30 -Washington New York 2010-10-30 07:00:00 -27.0 5904 2010-10-30 -Washington New York 2010-10-30 07:00:00 -16.0 5917 2010-10-30 -Baltimore New York 2010-10-22 07:00:00 -12.0 1064 2010-10-22 +Chicago New York 2010-10-22 07:00:00 -11.0 3198 2010-10-22 +Cleveland New York 2010-10-22 07:00:00 -3.0 3014 2010-10-22 Baltimore New York 2010-10-22 07:00:00 54.0 1142 2010-10-22 -Baltimore New York 2010-10-22 07:00:00 18.0 1599 2010-10-22 -Chicago New York 2010-10-22 07:00:00 2.0 361 2010-10-22 -Chicago New York 2010-10-22 07:00:00 24.0 897 2010-10-22 +Baltimore New York 2010-10-22 07:00:00 -12.0 1064 2010-10-22 +Cleveland New York 2010-10-22 07:00:00 -25.0 2646 2010-10-22 +Cleveland New York 2010-10-22 07:00:00 1.0 2630 2010-10-22 Chicago New York 2010-10-22 07:00:00 16.0 1531 2010-10-22 +Chicago New York 2010-10-22 07:00:00 24.0 897 2010-10-22 +Chicago New York 2010-10-22 07:00:00 2.0 361 2010-10-22 +Baltimore New York 2010-10-22 07:00:00 18.0 1599 2010-10-22 Chicago New York 2010-10-22 07:00:00 -6.0 1610 2010-10-22 -Chicago New York 2010-10-22 07:00:00 -11.0 3198 2010-10-22 -Cleveland New York 2010-10-22 07:00:00 1.0 2630 2010-10-22 -Cleveland New York 2010-10-22 07:00:00 -25.0 2646 2010-10-22 -Cleveland New York 2010-10-22 07:00:00 -3.0 3014 2010-10-22 -Baltimore New York 2010-10-31 07:00:00 -1.0 1599 2010-10-31 -Baltimore New York 2010-10-31 07:00:00 -14.0 2571 2010-10-31 Chicago New York 2010-10-31 07:00:00 -25.0 361 2010-10-31 Chicago New York 2010-10-31 07:00:00 -18.0 897 2010-10-31 Chicago New York 2010-10-31 07:00:00 -4.0 1531 2010-10-31 Chicago New York 2010-10-31 07:00:00 -22.0 1610 2010-10-31 Chicago New York 2010-10-31 07:00:00 -15.0 3198 2010-10-31 Washington New York 2010-10-31 07:00:00 -18.0 7282 2010-10-31 +Baltimore New York 2010-10-31 07:00:00 -1.0 1599 2010-10-31 +Baltimore New York 2010-10-31 07:00:00 -14.0 2571 2010-10-31 PREHOOK: query: explain vectorization expression select * from flights_tiny_parquet_partitioned_date sort by fl_num, fl_date limit 25 PREHOOK: type: QUERY @@ -4062,18 +4062,20 @@ POSTHOOK: Input: default@flights_tiny_parquet_partitioned_timestamp@fl_time=2010 POSTHOOK: Input: default@flights_tiny_parquet_partitioned_timestamp@fl_time=2010-10-30 07%3A00%3A00 POSTHOOK: Input: default@flights_tiny_parquet_partitioned_timestamp@fl_time=2010-10-31 07%3A00%3A00 #### A masked pattern was here #### +Cleveland New York 2010-10-28 -6.0 2646 2010-10-28 07:00:00 +Cleveland New York 2010-10-28 3.0 2630 2010-10-28 07:00:00 Baltimore New York 2010-10-28 -4.0 1064 2010-10-28 07:00:00 +Chicago New York 2010-10-28 3.0 1610 2010-10-28 07:00:00 +Washington New York 2010-10-28 45.0 7291 2010-10-28 07:00:00 Baltimore New York 2010-10-28 -14.0 1142 2010-10-28 07:00:00 Baltimore New York 2010-10-28 -14.0 1599 2010-10-28 07:00:00 Chicago New York 2010-10-28 2.0 361 2010-10-28 07:00:00 Chicago New York 2010-10-28 2.0 897 2010-10-28 07:00:00 Chicago New York 2010-10-28 -11.0 1531 2010-10-28 07:00:00 -Chicago New York 2010-10-28 3.0 1610 2010-10-28 07:00:00 Chicago New York 2010-10-28 -18.0 3198 2010-10-28 07:00:00 -Cleveland New York 2010-10-28 3.0 2630 2010-10-28 07:00:00 -Cleveland New York 2010-10-28 -6.0 2646 2010-10-28 07:00:00 Cleveland New York 2010-10-28 1.0 3014 2010-10-28 07:00:00 -Washington New York 2010-10-28 45.0 7291 2010-10-28 07:00:00 +Cleveland New York 2010-10-26 4.0 2630 2010-10-26 07:00:00 +Washington New York 2010-10-26 4.0 7291 2010-10-26 07:00:00 Baltimore New York 2010-10-26 -22.0 1064 2010-10-26 07:00:00 Baltimore New York 2010-10-26 123.0 1142 2010-10-26 07:00:00 Baltimore New York 2010-10-26 90.0 1599 2010-10-26 07:00:00 @@ -4081,48 +4083,51 @@ Chicago New York 2010-10-26 12.0 361 2010-10-26 07:00:00 Chicago New York 2010-10-26 0.0 897 2010-10-26 07:00:00 Chicago New York 2010-10-26 29.0 1531 2010-10-26 07:00:00 Chicago New York 2010-10-26 -17.0 1610 2010-10-26 07:00:00 -Chicago New York 2010-10-26 6.0 3198 2010-10-26 07:00:00 -Cleveland New York 2010-10-26 4.0 2630 2010-10-26 07:00:00 -Cleveland New York 2010-10-26 -27.0 2646 2010-10-26 07:00:00 -Cleveland New York 2010-10-26 -11.0 2662 2010-10-26 07:00:00 +Chicago New York 2010-10-26 6.0 3198 2010-10-26 07:00:00 Cleveland New York 2010-10-26 13.0 3014 2010-10-26 07:00:00 -Washington New York 2010-10-26 4.0 7291 2010-10-26 07:00:00 -Baltimore New York 2010-10-21 17.0 1064 2010-10-21 07:00:00 -Baltimore New York 2010-10-21 105.0 1142 2010-10-21 07:00:00 -Baltimore New York 2010-10-21 28.0 1599 2010-10-21 07:00:00 -Chicago New York 2010-10-21 142.0 361 2010-10-21 07:00:00 -Chicago New York 2010-10-21 77.0 897 2010-10-21 07:00:00 -Chicago New York 2010-10-21 53.0 1531 2010-10-21 07:00:00 +Cleveland New York 2010-10-26 -11.0 2662 2010-10-26 07:00:00 +Cleveland New York 2010-10-26 -27.0 2646 2010-10-26 07:00:00 Chicago New York 2010-10-21 -5.0 1610 2010-10-21 07:00:00 Chicago New York 2010-10-21 51.0 3198 2010-10-21 07:00:00 -Cleveland New York 2010-10-21 3.0 2630 2010-10-21 07:00:00 -Cleveland New York 2010-10-21 29.0 2646 2010-10-21 07:00:00 Cleveland New York 2010-10-21 72.0 3014 2010-10-21 07:00:00 +Cleveland New York 2010-10-21 29.0 2646 2010-10-21 07:00:00 +Baltimore New York 2010-10-21 17.0 1064 2010-10-21 07:00:00 +Cleveland New York 2010-10-21 3.0 2630 2010-10-21 07:00:00 +Baltimore New York 2010-10-21 28.0 1599 2010-10-21 07:00:00 +Chicago New York 2010-10-21 77.0 897 2010-10-21 07:00:00 Washington New York 2010-10-21 22.0 7291 2010-10-21 07:00:00 +Chicago New York 2010-10-21 142.0 361 2010-10-21 07:00:00 +Chicago New York 2010-10-21 53.0 1531 2010-10-21 07:00:00 +Baltimore New York 2010-10-21 105.0 1142 2010-10-21 07:00:00 +Washington New York 2010-10-23 -16.0 7274 2010-10-23 07:00:00 Baltimore New York 2010-10-23 18.0 272 2010-10-23 07:00:00 Baltimore New York 2010-10-23 -10.0 1805 2010-10-23 07:00:00 Baltimore New York 2010-10-23 6.0 3171 2010-10-23 07:00:00 -Chicago New York 2010-10-23 3.0 384 2010-10-23 07:00:00 -Chicago New York 2010-10-23 32.0 426 2010-10-23 07:00:00 -Chicago New York 2010-10-23 1.0 650 2010-10-23 07:00:00 -Chicago New York 2010-10-23 11.0 3085 2010-10-23 07:00:00 -Cleveland New York 2010-10-23 -21.0 2932 2010-10-23 07:00:00 -Washington New York 2010-10-23 -25.0 5832 2010-10-23 07:00:00 -Washington New York 2010-10-23 -21.0 5904 2010-10-23 07:00:00 Washington New York 2010-10-23 -18.0 5917 2010-10-23 07:00:00 -Washington New York 2010-10-23 -16.0 7274 2010-10-23 07:00:00 +Washington New York 2010-10-23 -21.0 5904 2010-10-23 07:00:00 +Washington New York 2010-10-23 -25.0 5832 2010-10-23 07:00:00 +Cleveland New York 2010-10-23 -21.0 2932 2010-10-23 07:00:00 +Chicago New York 2010-10-23 11.0 3085 2010-10-23 07:00:00 +Chicago New York 2010-10-23 1.0 650 2010-10-23 07:00:00 +Chicago New York 2010-10-23 32.0 426 2010-10-23 07:00:00 +Chicago New York 2010-10-23 3.0 384 2010-10-23 07:00:00 +Chicago New York 2010-10-24 -17.0 1610 2010-10-24 07:00:00 +Washington New York 2010-10-24 -26.0 7282 2010-10-24 07:00:00 Baltimore New York 2010-10-24 12.0 1599 2010-10-24 07:00:00 Baltimore New York 2010-10-24 20.0 2571 2010-10-24 07:00:00 -Chicago New York 2010-10-24 10.0 361 2010-10-24 07:00:00 Chicago New York 2010-10-24 113.0 897 2010-10-24 07:00:00 -Chicago New York 2010-10-24 -5.0 1531 2010-10-24 07:00:00 -Chicago New York 2010-10-24 -17.0 1610 2010-10-24 07:00:00 Chicago New York 2010-10-24 -3.0 3198 2010-10-24 07:00:00 Cleveland New York 2010-10-24 5.0 2254 2010-10-24 07:00:00 Cleveland New York 2010-10-24 -11.0 2630 2010-10-24 07:00:00 Cleveland New York 2010-10-24 -20.0 2646 2010-10-24 07:00:00 Cleveland New York 2010-10-24 -9.0 3014 2010-10-24 07:00:00 -Washington New York 2010-10-24 -26.0 7282 2010-10-24 07:00:00 +Chicago New York 2010-10-24 10.0 361 2010-10-24 07:00:00 +Chicago New York 2010-10-24 -5.0 1531 2010-10-24 07:00:00 +Chicago New York 2010-10-25 -16.0 3198 2010-10-25 07:00:00 +Cleveland New York 2010-10-25 81.0 2646 2010-10-25 07:00:00 +Cleveland New York 2010-10-25 -4.0 2630 2010-10-25 07:00:00 +Washington New York 2010-10-25 9.0 7291 2010-10-25 07:00:00 +Cleveland New York 2010-10-25 42.0 3014 2010-10-25 07:00:00 Baltimore New York 2010-10-25 -25.0 1064 2010-10-25 07:00:00 Baltimore New York 2010-10-25 92.0 1142 2010-10-25 07:00:00 Baltimore New York 2010-10-25 106.0 1599 2010-10-25 07:00:00 @@ -4130,46 +4135,40 @@ Chicago New York 2010-10-25 31.0 361 2010-10-25 07:00:00 Chicago New York 2010-10-25 -1.0 897 2010-10-25 07:00:00 Chicago New York 2010-10-25 43.0 1531 2010-10-25 07:00:00 Chicago New York 2010-10-25 6.0 1610 2010-10-25 07:00:00 -Chicago New York 2010-10-25 -16.0 3198 2010-10-25 07:00:00 -Cleveland New York 2010-10-25 -4.0 2630 2010-10-25 07:00:00 -Cleveland New York 2010-10-25 81.0 2646 2010-10-25 07:00:00 -Cleveland New York 2010-10-25 42.0 3014 2010-10-25 07:00:00 -Washington New York 2010-10-25 9.0 7291 2010-10-25 07:00:00 -Baltimore New York 2010-10-29 -24.0 1064 2010-10-29 07:00:00 Baltimore New York 2010-10-29 21.0 1142 2010-10-29 07:00:00 -Baltimore New York 2010-10-29 -2.0 1599 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -12.0 361 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -11.0 897 2010-10-29 07:00:00 -Chicago New York 2010-10-29 15.0 1531 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -18.0 1610 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -4.0 3198 2010-10-29 07:00:00 -Cleveland New York 2010-10-29 -4.0 2630 2010-10-29 07:00:00 -Cleveland New York 2010-10-29 -19.0 2646 2010-10-29 07:00:00 -Cleveland New York 2010-10-29 -12.0 3014 2010-10-29 07:00:00 Washington New York 2010-10-29 1.0 7291 2010-10-29 07:00:00 -Baltimore New York 2010-10-20 -30.0 1064 2010-10-20 07:00:00 +Cleveland New York 2010-10-29 -12.0 3014 2010-10-29 07:00:00 +Cleveland New York 2010-10-29 -19.0 2646 2010-10-29 07:00:00 +Cleveland New York 2010-10-29 -4.0 2630 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -4.0 3198 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -18.0 1610 2010-10-29 07:00:00 +Chicago New York 2010-10-29 15.0 1531 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -11.0 897 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -12.0 361 2010-10-29 07:00:00 +Baltimore New York 2010-10-29 -2.0 1599 2010-10-29 07:00:00 +Baltimore New York 2010-10-29 -24.0 1064 2010-10-29 07:00:00 +Chicago New York 2010-10-20 42.0 361 2010-10-20 07:00:00 +Washington New York 2010-10-20 -2.0 7291 2010-10-20 07:00:00 +Chicago New York 2010-10-20 -2.0 3198 2010-10-20 07:00:00 Baltimore New York 2010-10-20 23.0 1142 2010-10-20 07:00:00 Baltimore New York 2010-10-20 6.0 1599 2010-10-20 07:00:00 -Chicago New York 2010-10-20 42.0 361 2010-10-20 07:00:00 +Baltimore New York 2010-10-20 -30.0 1064 2010-10-20 07:00:00 +Cleveland New York 2010-10-20 -8.0 2630 2010-10-20 07:00:00 +Cleveland New York 2010-10-20 -15.0 3014 2010-10-20 07:00:00 Chicago New York 2010-10-20 24.0 897 2010-10-20 07:00:00 Chicago New York 2010-10-20 15.0 1531 2010-10-20 07:00:00 Chicago New York 2010-10-20 -6.0 1610 2010-10-20 07:00:00 -Chicago New York 2010-10-20 -2.0 3198 2010-10-20 07:00:00 -Cleveland New York 2010-10-20 -8.0 2630 2010-10-20 07:00:00 -Cleveland New York 2010-10-20 -15.0 3014 2010-10-20 07:00:00 -Washington New York 2010-10-20 -2.0 7291 2010-10-20 07:00:00 -Baltimore New York 2010-10-27 -18.0 1064 2010-10-27 07:00:00 -Baltimore New York 2010-10-27 49.0 1142 2010-10-27 07:00:00 -Baltimore New York 2010-10-27 92.0 1599 2010-10-27 07:00:00 -Chicago New York 2010-10-27 148.0 361 2010-10-27 07:00:00 -Chicago New York 2010-10-27 -11.0 897 2010-10-27 07:00:00 -Chicago New York 2010-10-27 70.0 1531 2010-10-27 07:00:00 -Chicago New York 2010-10-27 8.0 1610 2010-10-27 07:00:00 +Washington New York 2010-10-27 26.0 7291 2010-10-27 07:00:00 Chicago New York 2010-10-27 21.0 3198 2010-10-27 07:00:00 +Chicago New York 2010-10-27 8.0 1610 2010-10-27 07:00:00 Cleveland New York 2010-10-27 16.0 2630 2010-10-27 07:00:00 Cleveland New York 2010-10-27 27.0 3014 2010-10-27 07:00:00 -Washington New York 2010-10-27 26.0 7291 2010-10-27 07:00:00 -Baltimore New York 2010-10-30 14.0 272 2010-10-30 07:00:00 +Chicago New York 2010-10-27 -11.0 897 2010-10-27 07:00:00 +Chicago New York 2010-10-27 148.0 361 2010-10-27 07:00:00 +Baltimore New York 2010-10-27 92.0 1599 2010-10-27 07:00:00 +Baltimore New York 2010-10-27 49.0 1142 2010-10-27 07:00:00 +Baltimore New York 2010-10-27 -18.0 1064 2010-10-27 07:00:00 +Chicago New York 2010-10-27 70.0 1531 2010-10-27 07:00:00 Baltimore New York 2010-10-30 -1.0 1805 2010-10-30 07:00:00 Baltimore New York 2010-10-30 5.0 3171 2010-10-30 07:00:00 Chicago New York 2010-10-30 -6.0 384 2010-10-30 07:00:00 @@ -4180,17 +4179,19 @@ Cleveland New York 2010-10-30 -23.0 2018 2010-10-30 07:00:00 Cleveland New York 2010-10-30 -12.0 2932 2010-10-30 07:00:00 Washington New York 2010-10-30 -27.0 5904 2010-10-30 07:00:00 Washington New York 2010-10-30 -16.0 5917 2010-10-30 07:00:00 +Baltimore New York 2010-10-30 14.0 272 2010-10-30 07:00:00 +Cleveland New York 2010-10-22 1.0 2630 2010-10-22 07:00:00 Baltimore New York 2010-10-22 -12.0 1064 2010-10-22 07:00:00 -Baltimore New York 2010-10-22 54.0 1142 2010-10-22 07:00:00 -Baltimore New York 2010-10-22 18.0 1599 2010-10-22 07:00:00 -Chicago New York 2010-10-22 2.0 361 2010-10-22 07:00:00 -Chicago New York 2010-10-22 24.0 897 2010-10-22 07:00:00 -Chicago New York 2010-10-22 16.0 1531 2010-10-22 07:00:00 -Chicago New York 2010-10-22 -6.0 1610 2010-10-22 07:00:00 Chicago New York 2010-10-22 -11.0 3198 2010-10-22 07:00:00 -Cleveland New York 2010-10-22 1.0 2630 2010-10-22 07:00:00 -Cleveland New York 2010-10-22 -25.0 2646 2010-10-22 07:00:00 +Chicago New York 2010-10-22 -6.0 1610 2010-10-22 07:00:00 +Chicago New York 2010-10-22 16.0 1531 2010-10-22 07:00:00 +Chicago New York 2010-10-22 24.0 897 2010-10-22 07:00:00 +Chicago New York 2010-10-22 2.0 361 2010-10-22 07:00:00 +Baltimore New York 2010-10-22 18.0 1599 2010-10-22 07:00:00 +Baltimore New York 2010-10-22 54.0 1142 2010-10-22 07:00:00 Cleveland New York 2010-10-22 -3.0 3014 2010-10-22 07:00:00 +Cleveland New York 2010-10-22 -25.0 2646 2010-10-22 07:00:00 +Washington New York 2010-10-31 -18.0 7282 2010-10-31 07:00:00 Baltimore New York 2010-10-31 -1.0 1599 2010-10-31 07:00:00 Baltimore New York 2010-10-31 -14.0 2571 2010-10-31 07:00:00 Chicago New York 2010-10-31 -25.0 361 2010-10-31 07:00:00 @@ -4198,7 +4199,6 @@ Chicago New York 2010-10-31 -18.0 897 2010-10-31 07:00:00 Chicago New York 2010-10-31 -4.0 1531 2010-10-31 07:00:00 Chicago New York 2010-10-31 -22.0 1610 2010-10-31 07:00:00 Chicago New York 2010-10-31 -15.0 3198 2010-10-31 07:00:00 -Washington New York 2010-10-31 -18.0 7282 2010-10-31 07:00:00 PREHOOK: query: select * from flights_tiny_parquet_partitioned_timestamp sort by fl_num, fl_time limit 25 PREHOOK: type: QUERY PREHOOK: Input: default@flights_tiny_parquet_partitioned_timestamp @@ -4394,18 +4394,20 @@ POSTHOOK: Input: default@flights_tiny_parquet_partitioned_timestamp@fl_time=2010 POSTHOOK: Input: default@flights_tiny_parquet_partitioned_timestamp@fl_time=2010-10-30 07%3A00%3A00 POSTHOOK: Input: default@flights_tiny_parquet_partitioned_timestamp@fl_time=2010-10-31 07%3A00%3A00 #### A masked pattern was here #### +Cleveland New York 2010-10-28 -6.0 2646 2010-10-28 07:00:00 +Cleveland New York 2010-10-28 3.0 2630 2010-10-28 07:00:00 Baltimore New York 2010-10-28 -4.0 1064 2010-10-28 07:00:00 +Chicago New York 2010-10-28 3.0 1610 2010-10-28 07:00:00 +Washington New York 2010-10-28 45.0 7291 2010-10-28 07:00:00 Baltimore New York 2010-10-28 -14.0 1142 2010-10-28 07:00:00 Baltimore New York 2010-10-28 -14.0 1599 2010-10-28 07:00:00 Chicago New York 2010-10-28 2.0 361 2010-10-28 07:00:00 Chicago New York 2010-10-28 2.0 897 2010-10-28 07:00:00 Chicago New York 2010-10-28 -11.0 1531 2010-10-28 07:00:00 -Chicago New York 2010-10-28 3.0 1610 2010-10-28 07:00:00 Chicago New York 2010-10-28 -18.0 3198 2010-10-28 07:00:00 -Cleveland New York 2010-10-28 3.0 2630 2010-10-28 07:00:00 -Cleveland New York 2010-10-28 -6.0 2646 2010-10-28 07:00:00 Cleveland New York 2010-10-28 1.0 3014 2010-10-28 07:00:00 -Washington New York 2010-10-28 45.0 7291 2010-10-28 07:00:00 +Cleveland New York 2010-10-26 4.0 2630 2010-10-26 07:00:00 +Washington New York 2010-10-26 4.0 7291 2010-10-26 07:00:00 Baltimore New York 2010-10-26 -22.0 1064 2010-10-26 07:00:00 Baltimore New York 2010-10-26 123.0 1142 2010-10-26 07:00:00 Baltimore New York 2010-10-26 90.0 1599 2010-10-26 07:00:00 @@ -4414,47 +4416,50 @@ Chicago New York 2010-10-26 0.0 897 2010-10-26 07:00:00 Chicago New York 2010-10-26 29.0 1531 2010-10-26 07:00:00 Chicago New York 2010-10-26 -17.0 1610 2010-10-26 07:00:00 Chicago New York 2010-10-26 6.0 3198 2010-10-26 07:00:00 -Cleveland New York 2010-10-26 4.0 2630 2010-10-26 07:00:00 -Cleveland New York 2010-10-26 -27.0 2646 2010-10-26 07:00:00 -Cleveland New York 2010-10-26 -11.0 2662 2010-10-26 07:00:00 Cleveland New York 2010-10-26 13.0 3014 2010-10-26 07:00:00 -Washington New York 2010-10-26 4.0 7291 2010-10-26 07:00:00 -Baltimore New York 2010-10-21 17.0 1064 2010-10-21 07:00:00 -Baltimore New York 2010-10-21 105.0 1142 2010-10-21 07:00:00 -Baltimore New York 2010-10-21 28.0 1599 2010-10-21 07:00:00 -Chicago New York 2010-10-21 142.0 361 2010-10-21 07:00:00 -Chicago New York 2010-10-21 77.0 897 2010-10-21 07:00:00 -Chicago New York 2010-10-21 53.0 1531 2010-10-21 07:00:00 +Cleveland New York 2010-10-26 -11.0 2662 2010-10-26 07:00:00 +Cleveland New York 2010-10-26 -27.0 2646 2010-10-26 07:00:00 Chicago New York 2010-10-21 -5.0 1610 2010-10-21 07:00:00 Chicago New York 2010-10-21 51.0 3198 2010-10-21 07:00:00 -Cleveland New York 2010-10-21 3.0 2630 2010-10-21 07:00:00 -Cleveland New York 2010-10-21 29.0 2646 2010-10-21 07:00:00 Cleveland New York 2010-10-21 72.0 3014 2010-10-21 07:00:00 +Cleveland New York 2010-10-21 29.0 2646 2010-10-21 07:00:00 +Baltimore New York 2010-10-21 17.0 1064 2010-10-21 07:00:00 +Cleveland New York 2010-10-21 3.0 2630 2010-10-21 07:00:00 +Baltimore New York 2010-10-21 28.0 1599 2010-10-21 07:00:00 +Chicago New York 2010-10-21 77.0 897 2010-10-21 07:00:00 Washington New York 2010-10-21 22.0 7291 2010-10-21 07:00:00 +Chicago New York 2010-10-21 142.0 361 2010-10-21 07:00:00 +Chicago New York 2010-10-21 53.0 1531 2010-10-21 07:00:00 +Baltimore New York 2010-10-21 105.0 1142 2010-10-21 07:00:00 +Washington New York 2010-10-23 -16.0 7274 2010-10-23 07:00:00 Baltimore New York 2010-10-23 18.0 272 2010-10-23 07:00:00 Baltimore New York 2010-10-23 -10.0 1805 2010-10-23 07:00:00 Baltimore New York 2010-10-23 6.0 3171 2010-10-23 07:00:00 -Chicago New York 2010-10-23 3.0 384 2010-10-23 07:00:00 -Chicago New York 2010-10-23 32.0 426 2010-10-23 07:00:00 -Chicago New York 2010-10-23 1.0 650 2010-10-23 07:00:00 -Chicago New York 2010-10-23 11.0 3085 2010-10-23 07:00:00 -Cleveland New York 2010-10-23 -21.0 2932 2010-10-23 07:00:00 -Washington New York 2010-10-23 -25.0 5832 2010-10-23 07:00:00 -Washington New York 2010-10-23 -21.0 5904 2010-10-23 07:00:00 Washington New York 2010-10-23 -18.0 5917 2010-10-23 07:00:00 -Washington New York 2010-10-23 -16.0 7274 2010-10-23 07:00:00 +Washington New York 2010-10-23 -21.0 5904 2010-10-23 07:00:00 +Washington New York 2010-10-23 -25.0 5832 2010-10-23 07:00:00 +Cleveland New York 2010-10-23 -21.0 2932 2010-10-23 07:00:00 +Chicago New York 2010-10-23 11.0 3085 2010-10-23 07:00:00 +Chicago New York 2010-10-23 1.0 650 2010-10-23 07:00:00 +Chicago New York 2010-10-23 32.0 426 2010-10-23 07:00:00 +Chicago New York 2010-10-23 3.0 384 2010-10-23 07:00:00 +Chicago New York 2010-10-24 -17.0 1610 2010-10-24 07:00:00 +Washington New York 2010-10-24 -26.0 7282 2010-10-24 07:00:00 Baltimore New York 2010-10-24 12.0 1599 2010-10-24 07:00:00 Baltimore New York 2010-10-24 20.0 2571 2010-10-24 07:00:00 -Chicago New York 2010-10-24 10.0 361 2010-10-24 07:00:00 Chicago New York 2010-10-24 113.0 897 2010-10-24 07:00:00 -Chicago New York 2010-10-24 -5.0 1531 2010-10-24 07:00:00 -Chicago New York 2010-10-24 -17.0 1610 2010-10-24 07:00:00 Chicago New York 2010-10-24 -3.0 3198 2010-10-24 07:00:00 Cleveland New York 2010-10-24 5.0 2254 2010-10-24 07:00:00 Cleveland New York 2010-10-24 -11.0 2630 2010-10-24 07:00:00 Cleveland New York 2010-10-24 -20.0 2646 2010-10-24 07:00:00 Cleveland New York 2010-10-24 -9.0 3014 2010-10-24 07:00:00 -Washington New York 2010-10-24 -26.0 7282 2010-10-24 07:00:00 +Chicago New York 2010-10-24 10.0 361 2010-10-24 07:00:00 +Chicago New York 2010-10-24 -5.0 1531 2010-10-24 07:00:00 +Chicago New York 2010-10-25 -16.0 3198 2010-10-25 07:00:00 +Cleveland New York 2010-10-25 81.0 2646 2010-10-25 07:00:00 +Cleveland New York 2010-10-25 -4.0 2630 2010-10-25 07:00:00 +Washington New York 2010-10-25 9.0 7291 2010-10-25 07:00:00 +Cleveland New York 2010-10-25 42.0 3014 2010-10-25 07:00:00 Baltimore New York 2010-10-25 -25.0 1064 2010-10-25 07:00:00 Baltimore New York 2010-10-25 92.0 1142 2010-10-25 07:00:00 Baltimore New York 2010-10-25 106.0 1599 2010-10-25 07:00:00 @@ -4462,46 +4467,40 @@ Chicago New York 2010-10-25 31.0 361 2010-10-25 07:00:00 Chicago New York 2010-10-25 -1.0 897 2010-10-25 07:00:00 Chicago New York 2010-10-25 43.0 1531 2010-10-25 07:00:00 Chicago New York 2010-10-25 6.0 1610 2010-10-25 07:00:00 -Chicago New York 2010-10-25 -16.0 3198 2010-10-25 07:00:00 -Cleveland New York 2010-10-25 -4.0 2630 2010-10-25 07:00:00 -Cleveland New York 2010-10-25 81.0 2646 2010-10-25 07:00:00 -Cleveland New York 2010-10-25 42.0 3014 2010-10-25 07:00:00 -Washington New York 2010-10-25 9.0 7291 2010-10-25 07:00:00 -Baltimore New York 2010-10-29 -24.0 1064 2010-10-29 07:00:00 Baltimore New York 2010-10-29 21.0 1142 2010-10-29 07:00:00 -Baltimore New York 2010-10-29 -2.0 1599 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -12.0 361 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -11.0 897 2010-10-29 07:00:00 -Chicago New York 2010-10-29 15.0 1531 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -18.0 1610 2010-10-29 07:00:00 -Chicago New York 2010-10-29 -4.0 3198 2010-10-29 07:00:00 -Cleveland New York 2010-10-29 -4.0 2630 2010-10-29 07:00:00 -Cleveland New York 2010-10-29 -19.0 2646 2010-10-29 07:00:00 -Cleveland New York 2010-10-29 -12.0 3014 2010-10-29 07:00:00 Washington New York 2010-10-29 1.0 7291 2010-10-29 07:00:00 -Baltimore New York 2010-10-20 -30.0 1064 2010-10-20 07:00:00 +Cleveland New York 2010-10-29 -12.0 3014 2010-10-29 07:00:00 +Cleveland New York 2010-10-29 -19.0 2646 2010-10-29 07:00:00 +Cleveland New York 2010-10-29 -4.0 2630 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -4.0 3198 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -18.0 1610 2010-10-29 07:00:00 +Chicago New York 2010-10-29 15.0 1531 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -11.0 897 2010-10-29 07:00:00 +Chicago New York 2010-10-29 -12.0 361 2010-10-29 07:00:00 +Baltimore New York 2010-10-29 -2.0 1599 2010-10-29 07:00:00 +Baltimore New York 2010-10-29 -24.0 1064 2010-10-29 07:00:00 +Chicago New York 2010-10-20 42.0 361 2010-10-20 07:00:00 +Washington New York 2010-10-20 -2.0 7291 2010-10-20 07:00:00 +Chicago New York 2010-10-20 -2.0 3198 2010-10-20 07:00:00 Baltimore New York 2010-10-20 23.0 1142 2010-10-20 07:00:00 Baltimore New York 2010-10-20 6.0 1599 2010-10-20 07:00:00 -Chicago New York 2010-10-20 42.0 361 2010-10-20 07:00:00 +Baltimore New York 2010-10-20 -30.0 1064 2010-10-20 07:00:00 +Cleveland New York 2010-10-20 -8.0 2630 2010-10-20 07:00:00 +Cleveland New York 2010-10-20 -15.0 3014 2010-10-20 07:00:00 Chicago New York 2010-10-20 24.0 897 2010-10-20 07:00:00 Chicago New York 2010-10-20 15.0 1531 2010-10-20 07:00:00 Chicago New York 2010-10-20 -6.0 1610 2010-10-20 07:00:00 -Chicago New York 2010-10-20 -2.0 3198 2010-10-20 07:00:00 -Cleveland New York 2010-10-20 -8.0 2630 2010-10-20 07:00:00 -Cleveland New York 2010-10-20 -15.0 3014 2010-10-20 07:00:00 -Washington New York 2010-10-20 -2.0 7291 2010-10-20 07:00:00 -Baltimore New York 2010-10-27 -18.0 1064 2010-10-27 07:00:00 -Baltimore New York 2010-10-27 49.0 1142 2010-10-27 07:00:00 -Baltimore New York 2010-10-27 92.0 1599 2010-10-27 07:00:00 -Chicago New York 2010-10-27 148.0 361 2010-10-27 07:00:00 -Chicago New York 2010-10-27 -11.0 897 2010-10-27 07:00:00 -Chicago New York 2010-10-27 70.0 1531 2010-10-27 07:00:00 -Chicago New York 2010-10-27 8.0 1610 2010-10-27 07:00:00 +Washington New York 2010-10-27 26.0 7291 2010-10-27 07:00:00 Chicago New York 2010-10-27 21.0 3198 2010-10-27 07:00:00 +Chicago New York 2010-10-27 8.0 1610 2010-10-27 07:00:00 Cleveland New York 2010-10-27 16.0 2630 2010-10-27 07:00:00 Cleveland New York 2010-10-27 27.0 3014 2010-10-27 07:00:00 -Washington New York 2010-10-27 26.0 7291 2010-10-27 07:00:00 -Baltimore New York 2010-10-30 14.0 272 2010-10-30 07:00:00 +Chicago New York 2010-10-27 -11.0 897 2010-10-27 07:00:00 +Chicago New York 2010-10-27 148.0 361 2010-10-27 07:00:00 +Baltimore New York 2010-10-27 92.0 1599 2010-10-27 07:00:00 +Baltimore New York 2010-10-27 49.0 1142 2010-10-27 07:00:00 +Baltimore New York 2010-10-27 -18.0 1064 2010-10-27 07:00:00 +Chicago New York 2010-10-27 70.0 1531 2010-10-27 07:00:00 Baltimore New York 2010-10-30 -1.0 1805 2010-10-30 07:00:00 Baltimore New York 2010-10-30 5.0 3171 2010-10-30 07:00:00 Chicago New York 2010-10-30 -6.0 384 2010-10-30 07:00:00 @@ -4512,17 +4511,19 @@ Cleveland New York 2010-10-30 -23.0 2018 2010-10-30 07:00:00 Cleveland New York 2010-10-30 -12.0 2932 2010-10-30 07:00:00 Washington New York 2010-10-30 -27.0 5904 2010-10-30 07:00:00 Washington New York 2010-10-30 -16.0 5917 2010-10-30 07:00:00 +Baltimore New York 2010-10-30 14.0 272 2010-10-30 07:00:00 +Cleveland New York 2010-10-22 1.0 2630 2010-10-22 07:00:00 Baltimore New York 2010-10-22 -12.0 1064 2010-10-22 07:00:00 -Baltimore New York 2010-10-22 54.0 1142 2010-10-22 07:00:00 -Baltimore New York 2010-10-22 18.0 1599 2010-10-22 07:00:00 -Chicago New York 2010-10-22 2.0 361 2010-10-22 07:00:00 -Chicago New York 2010-10-22 24.0 897 2010-10-22 07:00:00 -Chicago New York 2010-10-22 16.0 1531 2010-10-22 07:00:00 -Chicago New York 2010-10-22 -6.0 1610 2010-10-22 07:00:00 Chicago New York 2010-10-22 -11.0 3198 2010-10-22 07:00:00 -Cleveland New York 2010-10-22 1.0 2630 2010-10-22 07:00:00 -Cleveland New York 2010-10-22 -25.0 2646 2010-10-22 07:00:00 +Chicago New York 2010-10-22 -6.0 1610 2010-10-22 07:00:00 +Chicago New York 2010-10-22 16.0 1531 2010-10-22 07:00:00 +Chicago New York 2010-10-22 24.0 897 2010-10-22 07:00:00 +Chicago New York 2010-10-22 2.0 361 2010-10-22 07:00:00 +Baltimore New York 2010-10-22 18.0 1599 2010-10-22 07:00:00 +Baltimore New York 2010-10-22 54.0 1142 2010-10-22 07:00:00 Cleveland New York 2010-10-22 -3.0 3014 2010-10-22 07:00:00 +Cleveland New York 2010-10-22 -25.0 2646 2010-10-22 07:00:00 +Washington New York 2010-10-31 -18.0 7282 2010-10-31 07:00:00 Baltimore New York 2010-10-31 -1.0 1599 2010-10-31 07:00:00 Baltimore New York 2010-10-31 -14.0 2571 2010-10-31 07:00:00 Chicago New York 2010-10-31 -25.0 361 2010-10-31 07:00:00 @@ -4530,7 +4531,6 @@ Chicago New York 2010-10-31 -18.0 897 2010-10-31 07:00:00 Chicago New York 2010-10-31 -4.0 1531 2010-10-31 07:00:00 Chicago New York 2010-10-31 -22.0 1610 2010-10-31 07:00:00 Chicago New York 2010-10-31 -15.0 3198 2010-10-31 07:00:00 -Washington New York 2010-10-31 -18.0 7282 2010-10-31 07:00:00 PREHOOK: query: explain vectorization expression select * from flights_tiny_parquet_partitioned_timestamp sort by fl_num, fl_time limit 25 PREHOOK: type: QUERY diff --git ql/src/test/results/clientpositive/llap_acid.q.out ql/src/test/results/clientpositive/llap_acid.q.out index cf092797adb9c31b195f24884c38273c6b81db1c..b96306c6aefd93a500ab827d4eae35a8c731b358 100644 --- ql/src/test/results/clientpositive/llap_acid.q.out +++ ql/src/test/results/clientpositive/llap_acid.q.out @@ -242,7 +242,7 @@ STAGE PLANS: neededVirtualColumns: [ROWID] partitionColumnCount: 1 partitionColumns: csmallint:smallint - scratchColumnTypeNames: [bigint] + scratchColumnTypeNames: [string] Reduce Vectorization: enabled: false enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true diff --git ql/src/test/results/clientpositive/llap_acid_fast.q.out ql/src/test/results/clientpositive/llap_acid_fast.q.out index 385e81d68c2b0147fb2f45e05a30a4d23f253437..d350c491bdbae3cd2025a2b142ea4b7d8a9b80bc 100644 --- ql/src/test/results/clientpositive/llap_acid_fast.q.out +++ ql/src/test/results/clientpositive/llap_acid_fast.q.out @@ -236,7 +236,7 @@ STAGE PLANS: neededVirtualColumns: [ROWID] partitionColumnCount: 1 partitionColumns: csmallint:smallint - scratchColumnTypeNames: [bigint] + scratchColumnTypeNames: [string] Reduce Vectorization: enabled: false enableConditionsMet: hive.vectorized.execution.reduce.enabled IS true diff --git ql/src/test/results/clientpositive/load_dyn_part1.q.out ql/src/test/results/clientpositive/load_dyn_part1.q.out index b43bd31128aed967f9f4f45bcbe03d07d704c13c..749de0c55d8b5b3219b0e21bd73ab43b04da5a6a 100644 --- ql/src/test/results/clientpositive/load_dyn_part1.q.out +++ ql/src/test/results/clientpositive/load_dyn_part1.q.out @@ -53,20 +53,11 @@ insert overwrite table nzhang_part2 partition(ds='2008-12-31', hr) select key, v POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-2 is a root stage - Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 - Stage-5 - Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 + Stage-0 depends on stages: Stage-2 Stage-3 depends on stages: Stage-0 - Stage-4 - Stage-6 - Stage-7 depends on stages: Stage-6 - Stage-14 depends on stages: Stage-2 , consists of Stage-11, Stage-10, Stage-12 - Stage-11 - Stage-1 depends on stages: Stage-11, Stage-10, Stage-13 - Stage-9 depends on stages: Stage-1 - Stage-10 - Stage-12 - Stage-13 depends on stages: Stage-12 + Stage-4 depends on stages: Stage-2 + Stage-1 depends on stages: Stage-4 + Stage-5 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-2 @@ -82,14 +73,12 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part1 + value expressions: _col0 (type: string), _col1 (type: string) Filter Operator predicate: (ds > '2008-04-08') (type: boolean) Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE @@ -99,21 +88,24 @@ STAGE PLANS: Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part2 - - Stage: Stage-8 - Conditional Operator - - Stage: Stage-5 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part1 Stage: Stage-0 Move Operator @@ -136,40 +128,26 @@ STAGE PLANS: Map Reduce Map Operator Tree: TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part1 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part1 - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-14 - Conditional Operator - - Stage: Stage-11 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part2 Stage: Stage-1 Move Operator @@ -184,40 +162,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part2 - Stage: Stage-9 + Stage: Stage-5 Stats Work Basic Stats Work: - Stage: Stage-10 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part2 - - Stage: Stage-12 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part2 - - Stage: Stage-13 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: from srcpart insert overwrite table nzhang_part1 partition (ds, hr) select key, value, ds, hr where ds <= '2008-04-08' insert overwrite table nzhang_part2 partition(ds='2008-12-31', hr) select key, value, hr where ds > '2008-04-08' diff --git ql/src/test/results/clientpositive/load_dyn_part10.q.out ql/src/test/results/clientpositive/load_dyn_part10.q.out index aea7798d0ba087b66a7d41d69481d924d8e59f0a..a840d7b11ea7c1888093685278475dbffa7a58da 100644 --- ql/src/test/results/clientpositive/load_dyn_part10.q.out +++ ql/src/test/results/clientpositive/load_dyn_part10.q.out @@ -57,14 +57,26 @@ STAGE PLANS: expressions: key (type: string), value (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part10 + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part10 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/load_dyn_part14.q.out ql/src/test/results/clientpositive/load_dyn_part14.q.out index 8e951b6e3e8bf6f4b93426cd8a489c0174e3ea8a..bc4aa19b103ef31999a8825cdda498523be93a69 100644 --- ql/src/test/results/clientpositive/load_dyn_part14.q.out +++ ql/src/test/results/clientpositive/load_dyn_part14.q.out @@ -44,16 +44,11 @@ select key, value from ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-2 depends on stages: Stage-1, Stage-9, Stage-10 - Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 - Stage-5 - Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 + Stage-2 depends on stages: Stage-1, Stage-4, Stage-5 + Stage-0 depends on stages: Stage-2 Stage-3 depends on stages: Stage-0 - Stage-4 - Stage-6 - Stage-7 depends on stages: Stage-6 - Stage-9 is a root stage - Stage-10 is a root stage + Stage-4 is a root stage + Stage-5 is a root stage STAGE PLANS: Stage: Stage-1 @@ -92,45 +87,44 @@ STAGE PLANS: TableScan Union Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col1 (type: string) + sort order: + + Map-reduce partition columns: _col1 (type: string) Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14 + value expressions: _col0 (type: string) TableScan Union Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col1 (type: string) + sort order: + + Map-reduce partition columns: _col1 (type: string) Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14 + value expressions: _col0 (type: string) TableScan Union Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col1 (type: string) + sort order: + + Map-reduce partition columns: _col1 (type: string) Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14 - - Stage: Stage-8 - Conditional Operator - - Stage: Stage-5 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + value expressions: _col0 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), KEY._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 941 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 6 Data size: 941 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part14 Stage: Stage-0 Move Operator @@ -152,36 +146,6 @@ STAGE PLANS: Map Reduce Map Operator Tree: TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14 - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-9 - Map Reduce - Map Operator Tree: - TableScan alias: src Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: COMPLETE Select Operator @@ -208,7 +172,7 @@ STAGE PLANS: output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Stage: Stage-10 + Stage: Stage-5 Map Reduce Map Operator Tree: TableScan diff --git ql/src/test/results/clientpositive/load_dyn_part3.q.out ql/src/test/results/clientpositive/load_dyn_part3.q.out index c63b4519f3ce56b099e42c254ec2699fdfe0ddb8..a1e7e9628d149090ca88bd77a5bc9a470278bf03 100644 --- ql/src/test/results/clientpositive/load_dyn_part3.q.out +++ ql/src/test/results/clientpositive/load_dyn_part3.q.out @@ -55,14 +55,26 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part3 + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part3 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/load_dyn_part4.q.out ql/src/test/results/clientpositive/load_dyn_part4.q.out index cf571921fcaec4da811d98f2c1097b08459d1d20..a632127a1c5510f2f4481e77e2e23abe4a79f1ac 100644 --- ql/src/test/results/clientpositive/load_dyn_part4.q.out +++ ql/src/test/results/clientpositive/load_dyn_part4.q.out @@ -65,14 +65,26 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part4 + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part4 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/load_dyn_part8.q.out ql/src/test/results/clientpositive/load_dyn_part8.q.out index 9e20fdc921ddde20abaf4c7357cf22c2acc86c5f..cdc2ddf22f0d4652546722bb01dd77b839e686c7 100644 --- ql/src/test/results/clientpositive/load_dyn_part8.q.out +++ ql/src/test/results/clientpositive/load_dyn_part8.q.out @@ -47,8 +47,9 @@ STAGE DEPENDENCIES: Stage-2 is a root stage Stage-0 depends on stages: Stage-2 Stage-3 depends on stages: Stage-0 - Stage-1 depends on stages: Stage-2 - Stage-4 depends on stages: Stage-1 + Stage-4 depends on stages: Stage-2 + Stage-1 depends on stages: Stage-4 + Stage-5 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-2 @@ -66,35 +67,15 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + null sort order: aa + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.nzhang_part8 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct nzhang_part8 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part8 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + tag: -1 + value expressions: _col0 (type: string), _col1 (type: string) + auto parallelism: false Filter Operator isSamplingPred: false predicate: (ds > '2008-04-08') (type: boolean) @@ -105,33 +86,21 @@ STAGE PLANS: Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - GlobalTableId: 2 + GlobalTableId: 0 #### A masked pattern was here #### NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-12-31/ - Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat properties: - bucket_count -1 column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.nzhang_part8 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct nzhang_part8 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part8 + columns _col0,_col1,_col2 + columns.types string,string,string + escape.delim \ + serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe TotalFiles: 1 - GatherStats: true + GatherStats: false MultiFileSpray: false Path -> Alias: #### A masked pattern was here #### @@ -333,6 +302,42 @@ STAGE PLANS: /srcpart/ds=2008-04-08/hr=12 [srcpart] /srcpart/ds=2008-04-09/hr=11 [srcpart] /srcpart/ds=2008-04-09/hr=12 [srcpart] + Needs Tagging: false + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + Dp Sort State: PARTITION_SORTED + NumFilesPerFileSink: 1 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.nzhang_part8 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct nzhang_part8 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part8 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Stage: Stage-0 Move Operator @@ -367,6 +372,85 @@ STAGE PLANS: Basic Stats Work: #### A masked pattern was here #### + Stage: Stage-4 + Map Reduce + Map Operator Tree: + TableScan + GatherStats: false + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: a + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: string), _col1 (type: string) + auto parallelism: false + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: -mr-10004 + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + column.name.delimiter , + columns _col0,_col1,_col2 + columns.types string,string,string + escape.delim \ + serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + properties: + column.name.delimiter , + columns _col0,_col1,_col2 + columns.types string,string,string + escape.delim \ + serialization.lib org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + Truncated Path -> Alias: +#### A masked pattern was here #### + Needs Tagging: false + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + GlobalTableId: 2 +#### A masked pattern was here #### + Dp Sort State: PARTITION_SORTED + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-12-31/ + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.nzhang_part8 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct nzhang_part8 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part8 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Stage: Stage-1 Move Operator tables: @@ -395,7 +479,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part8 - Stage: Stage-4 + Stage: Stage-5 Stats Work Basic Stats Work: #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/load_dyn_part9.q.out ql/src/test/results/clientpositive/load_dyn_part9.q.out index 77e689ca2d7c315e91bbe81a5f4063864c665a07..08ad3402163155c343b1ad38b000d99acb8382ea 100644 --- ql/src/test/results/clientpositive/load_dyn_part9.q.out +++ ql/src/test/results/clientpositive/load_dyn_part9.q.out @@ -57,14 +57,26 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part9 + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part9 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/merge3.q.out ql/src/test/results/clientpositive/merge3.q.out index 22a2820dd3348e6585b67e84ab2095f4e02f5d8e..02fe3d40535d84105c91b811e37f4575e94a823f 100644 --- ql/src/test/results/clientpositive/merge3.q.out +++ ql/src/test/results/clientpositive/merge3.q.out @@ -2392,35 +2392,15 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: a + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments - columns.types string:string -#### A masked pattern was here #### - name default.merge_src_part2 - partition_columns ds - partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_src_part2 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + tag: -1 + value expressions: _col0 (type: string), _col1 (type: string) + auto parallelism: false Path -> Alias: #### A masked pattern was here #### Path -> Partition: @@ -2440,7 +2420,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.merge_src_part - numFiles 2 + numFiles 1 numRows 1000 partition_columns ds partition_columns.types string @@ -2487,7 +2467,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.merge_src_part - numFiles 2 + numFiles 1 numRows 1000 partition_columns ds partition_columns.types string @@ -2521,6 +2501,42 @@ STAGE PLANS: Truncated Path -> Alias: /merge_src_part/ds=2008-04-08 [merge_src_part] /merge_src_part/ds=2008-04-09 [merge_src_part] + Needs Tagging: false + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + Dp Sort State: PARTITION_SORTED + NumFilesPerFileSink: 1 + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments + columns.types string:string +#### A masked pattern was here #### + name default.merge_src_part2 + partition_columns ds + partition_columns.types string + serialization.ddl struct merge_src_part2 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_src_part2 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Stage: Stage-7 Conditional Operator @@ -4815,8 +4831,9 @@ STAGE PLANS: outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator - null sort order: - sort order: + key expressions: _col2 (type: string) + null sort order: a + sort order: + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE tag: -1 @@ -4841,7 +4858,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.merge_src_part - numFiles 2 + numFiles 1 numRows 1000 partition_columns ds partition_columns.types string @@ -4888,7 +4905,7 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.merge_src_part - numFiles 2 + numFiles 1 numRows 1000 partition_columns ds partition_columns.types string @@ -4932,6 +4949,7 @@ STAGE PLANS: compressed: false GlobalTableId: 1 #### A masked pattern was here #### + Dp Sort State: PARTITION_SORTED NumFilesPerFileSink: 1 Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/merge4.q.out ql/src/test/results/clientpositive/merge4.q.out index de4c593d5302c85fa6fb38fcca55983e1f43a162..5c5d1c937a88ed13bafe4c02b0a1670973b4d469 100644 --- ql/src/test/results/clientpositive/merge4.q.out +++ ql/src/test/results/clientpositive/merge4.q.out @@ -33,14 +33,26 @@ STAGE PLANS: expressions: key (type: string), value (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part Stage: Stage-7 Conditional Operator @@ -2761,13 +2773,14 @@ select * from ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 - Stage-2 depends on stages: Stage-0 - Stage-3 + Stage-2 depends on stages: Stage-1 + Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 Stage-5 - Stage-6 depends on stages: Stage-5 + Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 + Stage-3 depends on stages: Stage-0 + Stage-4 + Stage-6 + Stage-7 depends on stages: Stage-6 STAGE PLANS: Stage: Stage-1 @@ -2817,17 +2830,40 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 259 Basic stats: COMPLETE Column stats: COMPLETE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 259 Basic stats: COMPLETE Column stats: COMPLETE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Stage: Stage-7 + Stage: Stage-2 + Map Reduce + Map Operator Tree: + TableScan + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 1 Data size: 259 Basic stats: COMPLETE Column stats: COMPLETE + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1 Data size: 259 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 1 Data size: 259 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part + + Stage: Stage-8 Conditional Operator - Stage: Stage-4 + Stage: Stage-5 Move Operator files: hdfs directory: true @@ -2846,11 +2882,11 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part - Stage: Stage-2 + Stage: Stage-3 Stats Work Basic Stats Work: - Stage: Stage-3 + Stage: Stage-4 Map Reduce Map Operator Tree: TableScan @@ -2862,7 +2898,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part - Stage: Stage-5 + Stage: Stage-6 Map Reduce Map Operator Tree: TableScan @@ -2874,7 +2910,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part - Stage: Stage-6 + Stage: Stage-7 Move Operator files: hdfs directory: true diff --git ql/src/test/results/clientpositive/merge_dynamic_partition3.q.out ql/src/test/results/clientpositive/merge_dynamic_partition3.q.out index 68d7bbe12237802eee56498d60ebce6f748e5608..f4ae6a935190327ed112a192214750eac552fa55 100644 --- ql/src/test/results/clientpositive/merge_dynamic_partition3.q.out +++ ql/src/test/results/clientpositive/merge_dynamic_partition3.q.out @@ -151,14 +151,26 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 174 Data size: 34830 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 174 Data size: 34830 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.merge_dynamic_part + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 174 Data size: 34830 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 174 Data size: 34830 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.merge_dynamic_part Stage: Stage-7 Conditional Operator @@ -274,9 +286,9 @@ outputformat:org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat columns:struct columns { string key, string value} partitioned:true partitionColumns:struct partition_columns { string ds, string hr} -totalNumberFiles:6 +totalNumberFiles:4 totalFileSize:34830 -maxFileSize:5812 -minFileSize:5791 +maxFileSize:11603 +minFileSize:5812 #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/merge_dynamic_partition4.q.out ql/src/test/results/clientpositive/merge_dynamic_partition4.q.out index 0da56cd9fca3bc66f57aa30202380caf5977314e..c4b3fc85a5b8d3a22b4602f7c5ea9e436402c4f9 100644 --- ql/src/test/results/clientpositive/merge_dynamic_partition4.q.out +++ ql/src/test/results/clientpositive/merge_dynamic_partition4.q.out @@ -152,14 +152,26 @@ STAGE PLANS: expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 2.0) = 0.0), 'a1', 'b1') (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 1000 Data size: 9624 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 1000 Data size: 9624 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.merge_dynamic_part + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 9624 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 1000 Data size: 9624 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.merge_dynamic_part Stage: Stage-7 Conditional Operator diff --git ql/src/test/results/clientpositive/merge_dynamic_partition5.q.out ql/src/test/results/clientpositive/merge_dynamic_partition5.q.out index 45237a15b94c9913acc4b3735ac828bde2ffcd76..7087e4859c584313b47341d40e74789a4f2885b0 100644 --- ql/src/test/results/clientpositive/merge_dynamic_partition5.q.out +++ ql/src/test/results/clientpositive/merge_dynamic_partition5.q.out @@ -128,14 +128,26 @@ STAGE PLANS: expressions: key (type: string), value (type: string), if(((UDFToDouble(key) % 100.0) = 0.0), 'a1', 'b1') (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 618 Data size: 5934 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 618 Data size: 5934 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.merge_dynamic_part + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 618 Data size: 5934 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 618 Data size: 5934 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.merge_dynamic_part Stage: Stage-7 Conditional Operator diff --git ql/src/test/results/clientpositive/orc_int_type_promotion.q.out ql/src/test/results/clientpositive/orc_int_type_promotion.q.out index d3837a3bd898007b686504c7cdc54ad33ca80084..4aabca6f86b809832cd31b88761fafbfd3feac5b 100644 --- ql/src/test/results/clientpositive/orc_int_type_promotion.q.out +++ ql/src/test/results/clientpositive/orc_int_type_promotion.q.out @@ -258,15 +258,15 @@ POSTHOOK: Input: default@src_part_orc@ds=2008-04-08 POSTHOOK: Input: default@src_part_orc@ds=2008-04-09 #### A masked pattern was here #### 238 val_238 2008-04-08 -86 val_86 2008-04-08 -311 val_311 2008-04-08 -27 val_27 2008-04-08 -165 val_165 2008-04-08 -409 val_409 2008-04-08 -255 val_255 2008-04-08 -278 val_278 2008-04-08 -98 val_98 2008-04-08 -484 val_484 2008-04-08 +97 val_97 2008-04-08 +200 val_200 2008-04-08 +400 val_400 2008-04-08 +403 val_403 2008-04-08 +169 val_169 2008-04-08 +90 val_90 2008-04-08 +126 val_126 2008-04-08 +222 val_222 2008-04-08 +477 val_477 2008-04-08 PREHOOK: query: alter table src_part_orc change key key bigint PREHOOK: type: ALTERTABLE_RENAMECOL PREHOOK: Input: default@src_part_orc @@ -288,12 +288,12 @@ POSTHOOK: Input: default@src_part_orc@ds=2008-04-08 POSTHOOK: Input: default@src_part_orc@ds=2008-04-09 #### A masked pattern was here #### 238 val_238 2008-04-08 -86 val_86 2008-04-08 -311 val_311 2008-04-08 -27 val_27 2008-04-08 -165 val_165 2008-04-08 -409 val_409 2008-04-08 -255 val_255 2008-04-08 -278 val_278 2008-04-08 -98 val_98 2008-04-08 -484 val_484 2008-04-08 +97 val_97 2008-04-08 +200 val_200 2008-04-08 +400 val_400 2008-04-08 +403 val_403 2008-04-08 +169 val_169 2008-04-08 +90 val_90 2008-04-08 +126 val_126 2008-04-08 +222 val_222 2008-04-08 +477 val_477 2008-04-08 diff --git ql/src/test/results/clientpositive/orc_merge2.q.out ql/src/test/results/clientpositive/orc_merge2.q.out index 5e8a1f0ccbadc1027b6f9ff2c0b8f64f6907e464..cbf1f9ea1451efe3863c9c308dbb928eff6aae0e 100644 --- ql/src/test/results/clientpositive/orc_merge2.q.out +++ ql/src/test/results/clientpositive/orc_merge2.q.out @@ -26,13 +26,8 @@ POSTHOOK: query: EXPLAIN INSERT OVERWRITE TABLE orcfile_merge2a PARTITION (one=' POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-0 depends on stages: Stage-1 Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 STAGE PLANS: Stage: Stage-1 @@ -45,23 +40,26 @@ STAGE PLANS: expressions: UDFToInteger(key) (type: int), value (type: string), (hash(key) pmod 10) (type: int), (hash(value) pmod 10) (type: int) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: int), _col3 (type: int) + sort order: ++ + Map-reduce partition columns: _col2 (type: int), _col3 (type: int) Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat - serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde - name: default.orcfile_merge2a - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + value expressions: _col0 (type: int), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: string), KEY._col2 (type: int), KEY._col3 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat + output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat + serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde + name: default.orcfile_merge2a Stage: Stage-0 Move Operator @@ -81,26 +79,6 @@ STAGE PLANS: Stats Work Basic Stats Work: - Stage: Stage-3 - Merge File Operator - Map Operator Tree: - ORC File Merge Operator - merge level: stripe - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - - Stage: Stage-5 - Merge File Operator - Map Operator Tree: - ORC File Merge Operator - merge level: stripe - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: INSERT OVERWRITE TABLE orcfile_merge2a PARTITION (one='1', two, three) SELECT key, value, PMOD(HASH(key), 10) as two, PMOD(HASH(value), 10) as three diff --git ql/src/test/results/clientpositive/spark/load_dyn_part1.q.out ql/src/test/results/clientpositive/spark/load_dyn_part1.q.out index 9c62fb344bc7778565cfb5f665e83e076d94f135..c3bfa2de01a099d351c9136bba793f1a5ff69f86 100644 --- ql/src/test/results/clientpositive/spark/load_dyn_part1.q.out +++ ql/src/test/results/clientpositive/spark/load_dyn_part1.q.out @@ -61,9 +61,12 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-2 Spark + Edges: + Reducer 2 <- Map 4 (PARTITION-LEVEL SORT, 2) + Reducer 3 <- Map 5 (PARTITION-LEVEL SORT, 2) #### A masked pattern was here #### Vertices: - Map 1 + Map 4 Map Operator Tree: TableScan alias: srcpart @@ -75,14 +78,17 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part1 + value expressions: _col0 (type: string), _col1 (type: string) + Map 5 + Map Operator Tree: + TableScan + alias: srcpart + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE Filter Operator predicate: (ds > '2008-04-08') (type: boolean) Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE @@ -90,14 +96,42 @@ STAGE PLANS: expressions: key (type: string), value (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part2 + value expressions: _col0 (type: string), _col1 (type: string) + Reducer 2 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part1 + Reducer 3 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part2 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/spark/load_dyn_part10.q.out ql/src/test/results/clientpositive/spark/load_dyn_part10.q.out index ad4668e14950ac980e7559e1219d4f2b7ddaae6f..304a786811cba75343a5f33d88de649f13efb759 100644 --- ql/src/test/results/clientpositive/spark/load_dyn_part10.q.out +++ ql/src/test/results/clientpositive/spark/load_dyn_part10.q.out @@ -49,6 +49,8 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 Spark + Edges: + Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 2) #### A masked pattern was here #### Vertices: Map 1 @@ -60,14 +62,27 @@ STAGE PLANS: expressions: key (type: string), value (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part10 + value expressions: _col0 (type: string), _col1 (type: string) + Reducer 2 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part10 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/spark/load_dyn_part14.q.out ql/src/test/results/clientpositive/spark/load_dyn_part14.q.out index 08d63506c01d7994e3d0e62b86a996c14e2ff3f1..f59a8ea5972a4d7f3269e546f3541066514a5529 100644 --- ql/src/test/results/clientpositive/spark/load_dyn_part14.q.out +++ ql/src/test/results/clientpositive/spark/load_dyn_part14.q.out @@ -52,8 +52,9 @@ STAGE PLANS: Spark Edges: Reducer 2 <- Map 1 (GROUP, 1) - Reducer 4 <- Map 1 (GROUP, 1) - Reducer 6 <- Map 1 (GROUP, 1) + Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 2), Reducer 5 (PARTITION-LEVEL SORT, 2), Reducer 7 (PARTITION-LEVEL SORT, 2) + Reducer 5 <- Map 1 (GROUP, 1) + Reducer 7 <- Map 1 (GROUP, 1) #### A masked pattern was here #### Vertices: Map 1 @@ -79,15 +80,28 @@ STAGE PLANS: expressions: 'k1' (type: string), null (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 2 Data size: 256 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col1 (type: string) + sort order: + + Map-reduce partition columns: _col1 (type: string) Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14 - Reducer 4 + value expressions: _col0 (type: string) + Reducer 3 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), KEY._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 6 Data size: 941 Basic stats: COMPLETE Column stats: COMPLETE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 6 Data size: 941 Basic stats: COMPLETE Column stats: COMPLETE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part14 + Reducer 5 Reduce Operator Tree: Limit Number of rows: 2 @@ -96,15 +110,13 @@ STAGE PLANS: expressions: 'k2' (type: string), '' (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 2 Data size: 340 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col1 (type: string) + sort order: + + Map-reduce partition columns: _col1 (type: string) Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14 - Reducer 6 + value expressions: _col0 (type: string) + Reducer 7 Reduce Operator Tree: Limit Number of rows: 2 @@ -113,14 +125,12 @@ STAGE PLANS: expressions: 'k3' (type: string), ' ' (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 2 Data size: 342 Basic stats: COMPLETE Column stats: COMPLETE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col1 (type: string) + sort order: + + Map-reduce partition columns: _col1 (type: string) Statistics: Num rows: 6 Data size: 938 Basic stats: COMPLETE Column stats: COMPLETE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part14 + value expressions: _col0 (type: string) Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/spark/load_dyn_part3.q.out ql/src/test/results/clientpositive/spark/load_dyn_part3.q.out index ca6729aba8d4133c31bd140c53e15fd81f5e1bd7..69c520677e2d6ccbbc182bd1b7d6e1b330218309 100644 --- ql/src/test/results/clientpositive/spark/load_dyn_part3.q.out +++ ql/src/test/results/clientpositive/spark/load_dyn_part3.q.out @@ -47,6 +47,8 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 Spark + Edges: + Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 2) #### A masked pattern was here #### Vertices: Map 1 @@ -58,14 +60,27 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part3 + value expressions: _col0 (type: string), _col1 (type: string) + Reducer 2 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part3 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/spark/load_dyn_part4.q.out ql/src/test/results/clientpositive/spark/load_dyn_part4.q.out index 368c08a501dc662844bf3ff9e8c12c7a6280d2a0..e7a2f4fe0c45d445144ce27a74b439d844085974 100644 --- ql/src/test/results/clientpositive/spark/load_dyn_part4.q.out +++ ql/src/test/results/clientpositive/spark/load_dyn_part4.q.out @@ -57,6 +57,8 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 Spark + Edges: + Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 2) #### A masked pattern was here #### Vertices: Map 1 @@ -68,14 +70,27 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part4 + value expressions: _col0 (type: string), _col1 (type: string) + Reducer 2 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part4 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/spark/load_dyn_part5.q.out ql/src/test/results/clientpositive/spark/load_dyn_part5.q.out index 4fc0d8c31ac78b855379ff9ce1ea0e9c31006f5c..25679366d8861b2b52feb8e464c343a39c2f5a63 100644 --- ql/src/test/results/clientpositive/spark/load_dyn_part5.q.out +++ ql/src/test/results/clientpositive/spark/load_dyn_part5.q.out @@ -34,6 +34,8 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 Spark + Edges: + Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 2) #### A masked pattern was here #### Vertices: Map 1 @@ -45,14 +47,27 @@ STAGE PLANS: expressions: key (type: string), value (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col1 (type: string) + sort order: + + Map-reduce partition columns: _col1 (type: string) Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part5 + value expressions: _col0 (type: string) + Reducer 2 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), KEY._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part5 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out index fea2d518acf00bcebd0e4871b3755aba33241158..ddb8c41041d2807f28803d8abcc4224cf7c2a97b 100644 --- ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out +++ ql/src/test/results/clientpositive/spark/load_dyn_part8.q.out @@ -53,9 +53,12 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-2 Spark + Edges: + Reducer 2 <- Map 4 (PARTITION-LEVEL SORT, 2) + Reducer 3 <- Map 5 (PARTITION-LEVEL SORT, 2) #### A masked pattern was here #### Vertices: - Map 1 + Map 4 Map Operator Tree: TableScan alias: srcpart @@ -69,35 +72,221 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - GlobalTableId: 1 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + null sort order: aa + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + tag: -1 + value expressions: _col0 (type: string), _col1 (type: string) + auto parallelism: false + Path -> Alias: +#### A masked pattern was here #### + Path -> Partition: +#### A masked pattern was here #### + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-08 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string #### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe #### A masked pattern was here #### - name default.nzhang_part8 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct nzhang_part8 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart #### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part8 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + Partition + base file name: hr=11 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-09 + hr 11 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart +#### A masked pattern was here #### + Partition + base file name: hr=12 + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + partition values: + ds 2008-04-09 + hr 12 + properties: + COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + numFiles 1 + numRows 500 + partition_columns ds/hr + partition_columns.types string:string + rawDataSize 5312 + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + totalSize 5812 +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.srcpart + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct srcpart { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.srcpart + name: default.srcpart + Truncated Path -> Alias: + /srcpart/ds=2008-04-08/hr=11 [srcpart] + /srcpart/ds=2008-04-08/hr=12 [srcpart] + /srcpart/ds=2008-04-09/hr=11 [srcpart] + /srcpart/ds=2008-04-09/hr=12 [srcpart] + Map 5 + Map Operator Tree: + TableScan + alias: srcpart + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + GatherStats: false Filter Operator isSamplingPred: false predicate: (ds > '2008-04-08') (type: boolean) @@ -106,36 +295,15 @@ STAGE PLANS: expressions: key (type: string), value (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - GlobalTableId: 2 -#### A masked pattern was here #### - NumFilesPerFileSink: 1 - Static Partition Specification: ds=2008-12-31/ + Reduce Output Operator + key expressions: _col2 (type: string) + null sort order: a + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE -#### A masked pattern was here #### - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - properties: - bucket_count -1 - column.name.delimiter , - columns key,value - columns.comments 'default','default' - columns.types string:string -#### A masked pattern was here #### - name default.nzhang_part8 - partition_columns ds/hr - partition_columns.types string:string - serialization.ddl struct nzhang_part8 { string key, string value} - serialization.format 1 - serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part8 - TotalFiles: 1 - GatherStats: true - MultiFileSpray: false + tag: -1 + value expressions: _col0 (type: string), _col1 (type: string) + auto parallelism: false Path -> Alias: #### A masked pattern was here #### Path -> Partition: @@ -336,6 +504,81 @@ STAGE PLANS: /srcpart/ds=2008-04-08/hr=12 [srcpart] /srcpart/ds=2008-04-09/hr=11 [srcpart] /srcpart/ds=2008-04-09/hr=12 [srcpart] + Reducer 2 + Needs Tagging: false + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + GlobalTableId: 1 +#### A masked pattern was here #### + Dp Sort State: PARTITION_SORTED + NumFilesPerFileSink: 1 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.nzhang_part8 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct nzhang_part8 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part8 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false + Reducer 3 + Needs Tagging: false + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + GlobalTableId: 2 +#### A masked pattern was here #### + Dp Sort State: PARTITION_SORTED + NumFilesPerFileSink: 1 + Static Partition Specification: ds=2008-12-31/ + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE +#### A masked pattern was here #### + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + properties: + bucket_count -1 + column.name.delimiter , + columns key,value + columns.comments 'default','default' + columns.types string:string +#### A masked pattern was here #### + name default.nzhang_part8 + partition_columns ds/hr + partition_columns.types string:string + serialization.ddl struct nzhang_part8 { string key, string value} + serialization.format 1 + serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe +#### A masked pattern was here #### + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part8 + TotalFiles: 1 + GatherStats: true + MultiFileSpray: false Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/spark/load_dyn_part9.q.out ql/src/test/results/clientpositive/spark/load_dyn_part9.q.out index 4bb08a05d90f447c4362610e9014c385703e5949..3d7f16506b9c63149960cb19dc3ccf0b17de68d7 100644 --- ql/src/test/results/clientpositive/spark/load_dyn_part9.q.out +++ ql/src/test/results/clientpositive/spark/load_dyn_part9.q.out @@ -49,6 +49,8 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 Spark + Edges: + Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 2) #### A masked pattern was here #### Vertices: Map 1 @@ -60,14 +62,27 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part9 + value expressions: _col0 (type: string), _col1 (type: string) + Reducer 2 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 1000 Data size: 10624 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part9 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/spark/orc_merge2.q.out ql/src/test/results/clientpositive/spark/orc_merge2.q.out index e05ba8c733f256a03efb603a7d8c0a0c8e4b7391..e224543be6626b88d3bed7500465b0d0ef504dd6 100644 --- ql/src/test/results/clientpositive/spark/orc_merge2.q.out +++ ql/src/test/results/clientpositive/spark/orc_merge2.q.out @@ -37,6 +37,8 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 Spark + Edges: + Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 4) #### A masked pattern was here #### Vertices: Map 1 @@ -48,14 +50,27 @@ STAGE PLANS: expressions: UDFToInteger(key) (type: int), value (type: string), (hash(key) pmod 10) (type: int), (hash(value) pmod 10) (type: int) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: int), _col3 (type: int) + sort order: ++ + Map-reduce partition columns: _col2 (type: int), _col3 (type: int) Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat - output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat - serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde - name: default.orcfile_merge2a + value expressions: _col0 (type: int), _col1 (type: string) + Reducer 2 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: int), VALUE._col1 (type: string), KEY._col2 (type: int), KEY._col3 (type: int) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 500 Data size: 5312 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat + output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat + serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde + name: default.orcfile_merge2a Stage: Stage-7 Conditional Operator diff --git ql/src/test/results/clientpositive/spark/stats2.q.out ql/src/test/results/clientpositive/spark/stats2.q.out index f609702e01b6a249f88272b6482827c3e5d46ac7..2cf462e1a04b2b53b1b52d885bbee0dce0476a19 100644 --- ql/src/test/results/clientpositive/spark/stats2.q.out +++ ql/src/test/results/clientpositive/spark/stats2.q.out @@ -19,6 +19,8 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 Spark + Edges: + Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 2) #### A masked pattern was here #### Vertices: Map 1 @@ -30,14 +32,27 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.analyze_t1 + value expressions: _col0 (type: string), _col1 (type: string) + Reducer 2 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.analyze_t1 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/spark/union_remove_15.q.out ql/src/test/results/clientpositive/spark/union_remove_15.q.out index 8a25ceba73f4bce89cdc1b950f60c2a81c1a7faa..8ecd7a7b45c3908ca3d5107292662577cecfc4d3 100644 --- ql/src/test/results/clientpositive/spark/union_remove_15.q.out +++ ql/src/test/results/clientpositive/spark/union_remove_15.q.out @@ -49,7 +49,8 @@ STAGE PLANS: Spark Edges: Reducer 2 <- Map 1 (GROUP, 2) - Reducer 4 <- Map 1 (GROUP, 2) + Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 2), Reducer 5 (PARTITION-LEVEL SORT, 2) + Reducer 5 <- Map 1 (GROUP, 2) #### A masked pattern was here #### Vertices: Map 1 @@ -85,15 +86,28 @@ STAGE PLANS: expressions: _col0 (type: string), _col1 (type: bigint), '1' (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.outputtbl1 - Reducer 4 + value expressions: _col0 (type: string), _col1 (type: bigint) + Reducer 3 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: bigint), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.outputtbl1 + Reducer 5 Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) @@ -105,14 +119,12 @@ STAGE PLANS: expressions: _col0 (type: string), _col1 (type: bigint), '2' (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.outputtbl1 + value expressions: _col0 (type: string), _col1 (type: bigint) Stage: Stage-0 Move Operator @@ -172,11 +184,11 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - numFiles 4 + numFiles 2 numPartitions 2 numRows 0 rawDataSize 0 - totalSize 332 + totalSize 178 #### A masked pattern was here #### # Storage Information diff --git ql/src/test/results/clientpositive/spark/union_remove_16.q.out ql/src/test/results/clientpositive/spark/union_remove_16.q.out index 8594301cbc92a9876abf6aa301d4fb120bcb1575..abc4376e7c779346b0d767fc49378c0ef5975e84 100644 --- ql/src/test/results/clientpositive/spark/union_remove_16.q.out +++ ql/src/test/results/clientpositive/spark/union_remove_16.q.out @@ -54,7 +54,8 @@ STAGE PLANS: Spark Edges: Reducer 2 <- Map 1 (GROUP, 2) - Reducer 4 <- Map 1 (GROUP, 2) + Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 2), Reducer 5 (PARTITION-LEVEL SORT, 2) + Reducer 5 <- Map 1 (GROUP, 2) #### A masked pattern was here #### Vertices: Map 1 @@ -90,15 +91,28 @@ STAGE PLANS: expressions: _col0 (type: string), _col1 (type: bigint), '1' (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.outputtbl1 - Reducer 4 + value expressions: _col0 (type: string), _col1 (type: bigint) + Reducer 3 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: bigint), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.outputtbl1 + Reducer 5 Reduce Operator Tree: Group By Operator aggregations: count(VALUE._col0) @@ -110,14 +124,12 @@ STAGE PLANS: expressions: _col0 (type: string), _col1 (type: bigint), '2' (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.outputtbl1 + value expressions: _col0 (type: string), _col1 (type: bigint) Stage: Stage-6 Conditional Operator @@ -214,11 +226,11 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - numFiles 4 + numFiles 2 numPartitions 2 numRows 0 rawDataSize 0 - totalSize 332 + totalSize 178 #### A masked pattern was here #### # Storage Information diff --git ql/src/test/results/clientpositive/spark/union_remove_17.q.out ql/src/test/results/clientpositive/spark/union_remove_17.q.out index cef4255535ef09008ee74872460a6f39b954dcbe..df95041a43e664fcb734a17c48f81ada94aac734 100644 --- ql/src/test/results/clientpositive/spark/union_remove_17.q.out +++ ql/src/test/results/clientpositive/spark/union_remove_17.q.out @@ -47,6 +47,8 @@ STAGE DEPENDENCIES: STAGE PLANS: Stage: Stage-1 Spark + Edges: + Reducer 2 <- Map 1 (PARTITION-LEVEL SORT, 2), Map 3 (PARTITION-LEVEL SORT, 2) #### A masked pattern was here #### Vertices: Map 1 @@ -62,15 +64,13 @@ STAGE PLANS: expressions: _col0 (type: string), UDFToLong(_col1) (type: bigint), _col2 (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.outputtbl1 - Map 2 + value expressions: _col0 (type: string), _col1 (type: bigint) + Map 3 Map Operator Tree: TableScan alias: inputtbl1 @@ -83,14 +83,27 @@ STAGE PLANS: expressions: _col0 (type: string), UDFToLong(_col1) (type: bigint), _col2 (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.outputtbl1 + value expressions: _col0 (type: string), _col1 (type: bigint) + Reducer 2 + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: bigint), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.outputtbl1 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/spark/union_remove_18.q.out ql/src/test/results/clientpositive/spark/union_remove_18.q.out index 51f2f163aa8b3bf7e4e3f268b5b191c663ed1a10..b276a937ecd65397644628d3039d9171889cbcee 100644 --- ql/src/test/results/clientpositive/spark/union_remove_18.q.out +++ ql/src/test/results/clientpositive/spark/union_remove_18.q.out @@ -49,7 +49,7 @@ STAGE PLANS: Spark Edges: Reducer 2 <- Map 1 (GROUP, 2) - Reducer 4 <- Map 1 (GROUP, 2) + Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 2), Reducer 2 (PARTITION-LEVEL SORT, 2) #### A masked pattern was here #### Vertices: Map 1 @@ -85,34 +85,27 @@ STAGE PLANS: expressions: _col0 (type: string), _col2 (type: bigint), _col1 (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1 - Reducer 4 + value expressions: _col0 (type: string), _col1 (type: bigint) + Reducer 3 Reduce Operator Tree: - Group By Operator - aggregations: count(VALUE._col0) - keys: KEY._col0 (type: string), KEY._col1 (type: string) - mode: mergepartial + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: bigint), KEY._col2 (type: string) outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), _col2 (type: bigint), _col1 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1 + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1 Stage: Stage-0 Move Operator @@ -184,7 +177,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - numFiles 12 + numFiles 6 numPartitions 6 numRows 0 rawDataSize 0 diff --git ql/src/test/results/clientpositive/spark/union_remove_25.q.out ql/src/test/results/clientpositive/spark/union_remove_25.q.out index f681428785c1bff7c1a91997c3609c051f3fd2b3..84a9f6d4bda041c067fffcd4d8c47176bab2278f 100644 --- ql/src/test/results/clientpositive/spark/union_remove_25.q.out +++ ql/src/test/results/clientpositive/spark/union_remove_25.q.out @@ -404,7 +404,7 @@ STAGE PLANS: Spark Edges: Reducer 2 <- Map 1 (GROUP, 1) - Reducer 4 <- Map 1 (GROUP, 1) + Reducer 3 <- Reducer 2 (PARTITION-LEVEL SORT, 2), Reducer 2 (PARTITION-LEVEL SORT, 2) #### A masked pattern was here #### Vertices: Map 1 @@ -437,35 +437,27 @@ STAGE PLANS: expressions: _col0 (type: string), UDFToLong(_col1) (type: bigint), '2008-04-08' (type: string), _col2 (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3 - Reducer 4 + value expressions: _col0 (type: string), _col1 (type: bigint) + Reducer 3 Reduce Operator Tree: Select Operator - expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), VALUE._col2 (type: string) - outputColumnNames: _col0, _col1, _col2 - Statistics: Num rows: 1000 Data size: 10000 Basic stats: COMPLETE Column stats: NONE - Limit - Number of rows: 1000 - Statistics: Num rows: 1000 Data size: 10000 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), UDFToLong(_col1) (type: bigint), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3 + expressions: VALUE._col0 (type: string), VALUE._col1 (type: bigint), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl3 Stage: Stage-0 Move Operator @@ -538,7 +530,7 @@ Database: default Table: outputtbl3 #### A masked pattern was here #### Partition Parameters: - numFiles 2 + numFiles 1 totalSize 6812 #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/stats2.q.out ql/src/test/results/clientpositive/stats2.q.out index 65691d319c3a9425b0011e2031ca0935e574adef..42c91d092c896b1835d68cb8ee905bb885ad86e1 100644 --- ql/src/test/results/clientpositive/stats2.q.out +++ ql/src/test/results/clientpositive/stats2.q.out @@ -27,14 +27,26 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.analyze_t1 + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2000 Data size: 21248 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.analyze_t1 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/stats4.q.out ql/src/test/results/clientpositive/stats4.q.out index 741e47b9480713b2c7b4d292b07fca3c6d31a88f..c3544f265ffc713c792cafade31254d750dff267 100644 --- ql/src/test/results/clientpositive/stats4.q.out +++ ql/src/test/results/clientpositive/stats4.q.out @@ -44,20 +44,11 @@ insert overwrite table nzhang_part2 partition(ds='2008-12-31', hr) select key, v POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-2 is a root stage - Stage-8 depends on stages: Stage-2 , consists of Stage-5, Stage-4, Stage-6 - Stage-5 - Stage-0 depends on stages: Stage-5, Stage-4, Stage-7 + Stage-0 depends on stages: Stage-2 Stage-3 depends on stages: Stage-0 - Stage-4 - Stage-6 - Stage-7 depends on stages: Stage-6 - Stage-14 depends on stages: Stage-2 , consists of Stage-11, Stage-10, Stage-12 - Stage-11 - Stage-1 depends on stages: Stage-11, Stage-10, Stage-13 - Stage-9 depends on stages: Stage-1 - Stage-10 - Stage-12 - Stage-13 depends on stages: Stage-12 + Stage-4 depends on stages: Stage-2 + Stage-1 depends on stages: Stage-4 + Stage-5 depends on stages: Stage-1 STAGE PLANS: Stage: Stage-2 @@ -73,14 +64,12 @@ STAGE PLANS: expressions: key (type: string), value (type: string), ds (type: string), hr (type: string) outputColumnNames: _col0, _col1, _col2, _col3 Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part1 + value expressions: _col0 (type: string), _col1 (type: string) Filter Operator predicate: (ds > '2008-04-08') (type: boolean) Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE @@ -90,21 +79,24 @@ STAGE PLANS: Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part2 - - Stage: Stage-8 - Conditional Operator - - Stage: Stage-5 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part1 Stage: Stage-0 Move Operator @@ -127,40 +119,26 @@ STAGE PLANS: Map Reduce Map Operator Tree: TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part1 - - Stage: Stage-6 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part1 - - Stage: Stage-7 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - - Stage: Stage-14 - Conditional Operator - - Stage: Stage-11 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: string), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 666 Data size: 7075 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.nzhang_part2 Stage: Stage-1 Move Operator @@ -175,40 +153,10 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part2 - Stage: Stage-9 + Stage: Stage-5 Stats Work Basic Stats Work: - Stage: Stage-10 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part2 - - Stage: Stage-12 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.nzhang_part2 - - Stage: Stage-13 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: from srcpart insert overwrite table nzhang_part1 partition (ds, hr) select key, value, ds, hr where ds <= '2008-04-08' insert overwrite table nzhang_part2 partition(ds='2008-12-31', hr) select key, value, hr where ds > '2008-04-08' diff --git ql/src/test/results/clientpositive/stats_empty_dyn_part.q.out ql/src/test/results/clientpositive/stats_empty_dyn_part.q.out index 54566546bb4745bd94ef3a82c1a4e3a7a233ba92..faaef749f688f6f16513f5bc82f0caf771223e8d 100644 --- ql/src/test/results/clientpositive/stats_empty_dyn_part.q.out +++ ql/src/test/results/clientpositive/stats_empty_dyn_part.q.out @@ -12,13 +12,8 @@ POSTHOOK: query: explain insert overwrite table tmptable partition (part) select POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-7 depends on stages: Stage-1 , consists of Stage-4, Stage-3, Stage-5 - Stage-4 - Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-0 depends on stages: Stage-1 Stage-2 depends on stages: Stage-0 - Stage-3 - Stage-5 - Stage-6 depends on stages: Stage-5 STAGE PLANS: Stage: Stage-1 @@ -34,23 +29,26 @@ STAGE PLANS: expressions: 'no_such_value' (type: string), value (type: string) outputColumnNames: _col0, _col1 Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col1 (type: string) + sort order: + + Map-reduce partition columns: _col1 (type: string) Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.tmptable - - Stage: Stage-7 - Conditional Operator - - Stage: Stage-4 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### + value expressions: _col0 (type: string) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), KEY._col1 (type: string) + outputColumnNames: _col0, _col1 + Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 250 Data size: 2656 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.tmptable Stage: Stage-0 Move Operator @@ -68,36 +66,6 @@ STAGE PLANS: Stats Work Basic Stats Work: - Stage: Stage-3 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.tmptable - - Stage: Stage-5 - Map Reduce - Map Operator Tree: - TableScan - File Output Operator - compressed: false - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.tmptable - - Stage: Stage-6 - Move Operator - files: - hdfs directory: true -#### A masked pattern was here #### - PREHOOK: query: insert overwrite table tmptable partition (part) select key, value from src where key = 'no_such_value' PREHOOK: type: QUERY PREHOOK: Input: default@src diff --git ql/src/test/results/clientpositive/union_remove_15.q.out ql/src/test/results/clientpositive/union_remove_15.q.out index 2934c4c0e9c2bf878d3427bd42e17e4ab5a5891b..5f37c10bc9e05debe3d046e76a595fa73ab081ab 100644 --- ql/src/test/results/clientpositive/union_remove_15.q.out +++ ql/src/test/results/clientpositive/union_remove_15.q.out @@ -42,8 +42,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-2 depends on stages: Stage-1, Stage-3 + Stage-0 depends on stages: Stage-2 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -81,12 +82,46 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.outputtbl1 + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + Stage: Stage-2 + Map Reduce + Map Operator Tree: + TableScan + Union + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: bigint) + TableScan + Union + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: bigint) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: bigint), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.outputtbl1 Stage: Stage-0 Move Operator @@ -100,7 +135,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-2 + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -135,12 +170,10 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.outputtbl1 + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe PREHOOK: query: insert overwrite table outputTbl1 partition (ds) SELECT * diff --git ql/src/test/results/clientpositive/union_remove_16.q.out ql/src/test/results/clientpositive/union_remove_16.q.out index 8be5ddd51162dbf691e0112fb314b0d695e957a6..b196c29f7bdf7800b987c21e6ed0bbae548b4b9f 100644 --- ql/src/test/results/clientpositive/union_remove_16.q.out +++ ql/src/test/results/clientpositive/union_remove_16.q.out @@ -42,13 +42,14 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-6 depends on stages: Stage-1, Stage-7 , consists of Stage-3, Stage-2, Stage-4 - Stage-3 - Stage-0 depends on stages: Stage-3, Stage-2, Stage-5 - Stage-2 + Stage-2 depends on stages: Stage-1, Stage-8 + Stage-7 depends on stages: Stage-2 , consists of Stage-4, Stage-3, Stage-5 Stage-4 - Stage-5 depends on stages: Stage-4 - Stage-7 is a root stage + Stage-0 depends on stages: Stage-4, Stage-3, Stage-6 + Stage-3 + Stage-5 + Stage-6 depends on stages: Stage-5 + Stage-8 is a root stage STAGE PLANS: Stage: Stage-1 @@ -86,17 +87,51 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.outputtbl1 + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe - Stage: Stage-6 + Stage: Stage-2 + Map Reduce + Map Operator Tree: + TableScan + Union + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: bigint) + TableScan + Union + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: bigint) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: bigint), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.outputtbl1 + + Stage: Stage-7 Conditional Operator - Stage: Stage-3 + Stage: Stage-4 Move Operator files: hdfs directory: true @@ -114,27 +149,27 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.outputtbl1 - Stage: Stage-2 + Stage: Stage-3 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-4 + Stage: Stage-5 Merge File Operator Map Operator Tree: RCFile Merge Operator merge level: block input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - Stage: Stage-5 + Stage: Stage-6 Move Operator files: hdfs directory: true #### A masked pattern was here #### - Stage: Stage-7 + Stage: Stage-8 Map Reduce Map Operator Tree: TableScan @@ -169,12 +204,10 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.outputtbl1 + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe PREHOOK: query: insert overwrite table outputTbl1 partition (ds) SELECT * diff --git ql/src/test/results/clientpositive/union_remove_17.q.out ql/src/test/results/clientpositive/union_remove_17.q.out index 4d35596b02e72c232cb5a1a718ef5f381cc8bc64..a05aaaa84df6ee28d9c7a64ba1786d3fc060d7f7 100644 --- ql/src/test/results/clientpositive/union_remove_17.q.out +++ ql/src/test/results/clientpositive/union_remove_17.q.out @@ -61,14 +61,12 @@ STAGE PLANS: expressions: _col0 (type: string), UDFToLong(_col1) (type: bigint), _col2 (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.outputtbl1 + value expressions: _col0 (type: string), _col1 (type: bigint) TableScan alias: inputtbl1 Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE @@ -82,14 +80,26 @@ STAGE PLANS: expressions: _col0 (type: string), UDFToLong(_col1) (type: bigint), _col2 (type: string) outputColumnNames: _col0, _col1, _col2 Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat - output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat - serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - name: default.outputtbl1 + value expressions: _col0 (type: string), _col1 (type: bigint) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: bigint), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat + output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat + serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe + name: default.outputtbl1 Stage: Stage-0 Move Operator diff --git ql/src/test/results/clientpositive/union_remove_18.q.out ql/src/test/results/clientpositive/union_remove_18.q.out index 3e181e31e1af341d456b669922996f59abd688a6..0996441f62a33222e81d3411ebac5b9ae4911ed3 100644 --- ql/src/test/results/clientpositive/union_remove_18.q.out +++ ql/src/test/results/clientpositive/union_remove_18.q.out @@ -42,8 +42,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-2 depends on stages: Stage-1, Stage-3 + Stage-0 depends on stages: Stage-2 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -81,12 +82,46 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1 + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + Stage: Stage-2 + Map Reduce + Map Operator Tree: + TableScan + Union + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: bigint) + TableScan + Union + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string) + sort order: + + Map-reduce partition columns: _col2 (type: string) + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: bigint) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: bigint), KEY._col2 (type: string) + outputColumnNames: _col0, _col1, _col2 + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2 Data size: 60 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl1 Stage: Stage-0 Move Operator @@ -100,7 +135,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1 - Stage: Stage-2 + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -135,12 +170,10 @@ STAGE PLANS: Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 30 Basic stats: COMPLETE Column stats: NONE table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl1 + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe PREHOOK: query: insert overwrite table outputTbl1 partition (ds) SELECT * @@ -200,7 +233,7 @@ Retention: 0 #### A masked pattern was here #### Table Type: MANAGED_TABLE Table Parameters: - numFiles 12 + numFiles 6 numPartitions 6 numRows 0 rawDataSize 0 diff --git ql/src/test/results/clientpositive/union_remove_25.q.out ql/src/test/results/clientpositive/union_remove_25.q.out index 20ab809cb1244617d03767a6f718c82cea1215a2..0634bc62d037af1dec4e489ccb64fe8ccdf16408 100644 --- ql/src/test/results/clientpositive/union_remove_25.q.out +++ ql/src/test/results/clientpositive/union_remove_25.q.out @@ -425,8 +425,9 @@ FROM ( POSTHOOK: type: QUERY STAGE DEPENDENCIES: Stage-1 is a root stage - Stage-0 depends on stages: Stage-1, Stage-2 - Stage-2 is a root stage + Stage-2 depends on stages: Stage-1, Stage-3 + Stage-0 depends on stages: Stage-2 + Stage-3 is a root stage STAGE PLANS: Stage: Stage-1 @@ -455,18 +456,56 @@ STAGE PLANS: Limit Number of rows: 1000 Statistics: Num rows: 1000 Data size: 10000 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), UDFToLong(_col1) (type: bigint), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1000 Data size: 10000 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1000 Data size: 10000 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3 + File Output Operator + compressed: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe + + Stage: Stage-2 + Map Reduce + Map Operator Tree: + TableScan + Union + Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), UDFToLong(_col1) (type: bigint), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) + Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: bigint) + TableScan + Union + Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE + Select Operator + expressions: _col0 (type: string), UDFToLong(_col1) (type: bigint), '2008-04-08' (type: string), _col2 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE + Reduce Output Operator + key expressions: _col2 (type: string), _col3 (type: string) + sort order: ++ + Map-reduce partition columns: _col2 (type: string), _col3 (type: string) + Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE + value expressions: _col0 (type: string), _col1 (type: bigint) + Reduce Operator Tree: + Select Operator + expressions: VALUE._col0 (type: string), VALUE._col1 (type: bigint), KEY._col2 (type: string), KEY._col3 (type: string) + outputColumnNames: _col0, _col1, _col2, _col3 + Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE + File Output Operator + compressed: false + Dp Sort State: PARTITION_SORTED + Statistics: Num rows: 2000 Data size: 20000 Basic stats: COMPLETE Column stats: NONE + table: + input format: org.apache.hadoop.mapred.TextInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat + serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe + name: default.outputtbl3 Stage: Stage-0 Move Operator @@ -481,7 +520,7 @@ STAGE PLANS: serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl3 - Stage: Stage-2 + Stage: Stage-3 Map Reduce Map Operator Tree: TableScan @@ -507,18 +546,12 @@ STAGE PLANS: Limit Number of rows: 1000 Statistics: Num rows: 1000 Data size: 10000 Basic stats: COMPLETE Column stats: NONE - Select Operator - expressions: _col0 (type: string), UDFToLong(_col1) (type: bigint), '2008-04-08' (type: string), _col2 (type: string) - outputColumnNames: _col0, _col1, _col2, _col3 - Statistics: Num rows: 1000 Data size: 10000 Basic stats: COMPLETE Column stats: NONE - File Output Operator - compressed: false - Statistics: Num rows: 1000 Data size: 10000 Basic stats: COMPLETE Column stats: NONE - table: - input format: org.apache.hadoop.mapred.TextInputFormat - output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat - serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - name: default.outputtbl3 + File Output Operator + compressed: false + table: + input format: org.apache.hadoop.mapred.SequenceFileInputFormat + output format: org.apache.hadoop.hive.ql.io.HiveSequenceFileOutputFormat + serde: org.apache.hadoop.hive.serde2.lazybinary.LazyBinarySerDe PREHOOK: query: insert overwrite table outputTbl3 partition(ds, hr) SELECT * @@ -578,7 +611,7 @@ Database: default Table: outputtbl3 #### A masked pattern was here #### Partition Parameters: - numFiles 2 + numFiles 1 totalSize 6812 #### A masked pattern was here #### diff --git ql/src/test/results/clientpositive/update_all_partitioned.q.out ql/src/test/results/clientpositive/update_all_partitioned.q.out index 4ee1e2e4cf5d5b056a6e5b0dc38fe39b35d26c92..7301f5990215011dda6dcd0bce2d33e67486a488 100644 --- ql/src/test/results/clientpositive/update_all_partitioned.q.out +++ ql/src/test/results/clientpositive/update_all_partitioned.q.out @@ -84,23 +84,29 @@ POSTHOOK: Input: default@acid_uap POSTHOOK: Input: default@acid_uap@ds=today POSTHOOK: Input: default@acid_uap@ds=tomorrow #### A masked pattern was here #### --1073279343 fred today --1073051226 fred today +-1073279343 oj1YrV5Wa today +-1073051226 A34p7oRr2WvUJNf today +-1072910839 0iqrc5 today -1072910839 fred today --1072081801 fred today +-1072081801 dPkN74F7 today -1072076362 fred today --1071480828 fred today +-1071480828 aw724t8c5558x2xneC624 today +-1071363017 Anj0oF today -1071363017 fred today --1070883071 fred today +-1070883071 0ruyd6Y50JpdGRf6HqD today -1070551679 fred today --1069736047 fred today -762 fred tomorrow -762 fred tomorrow +-1070551679 iUR3Q today +-1069736047 k17Am8uPHWk02cEf1jet today +762 40ks5556SV tomorrow +762 a10E76jX35YwquKCTA tomorrow 762 fred tomorrow 762 fred tomorrow 762 fred tomorrow -6981 fred tomorrow -6981 fred tomorrow +6981 NULL tomorrow +6981 1FNNhmiFLGw425NA13g tomorrow +6981 4KhrrQ0nJ7bMNTvhSCA tomorrow +6981 K630vaVf tomorrow +6981 Y5x3JuI3M8jngv5N tomorrow 6981 fred tomorrow 6981 fred tomorrow 6981 fred tomorrow diff --git standalone-metastore/src/gen/org/apache/hadoop/hive/metastore/annotation/package-info.java standalone-metastore/src/gen/org/apache/hadoop/hive/metastore/annotation/package-info.java deleted file mode 100644 index 0c7625d7846d56087c53e99ce671ef594831885d..0000000000000000000000000000000000000000 --- standalone-metastore/src/gen/org/apache/hadoop/hive/metastore/annotation/package-info.java +++ /dev/null @@ -1,8 +0,0 @@ -/* - * Generated by saveVersion.sh - */ -@MetastoreVersionAnnotation(version="3.0.0-SNAPSHOT", shortVersion="3.0.0", - revision="56f5e5957ea48404ae119721709d05600b59eeb1", branch="master", - user="sergey", date="Fri Nov 3 11:25:46 PDT 2017", url="git://HW10895.hsd1.wa.comcast.net/Users/sergey/git/hivegit2", - srcChecksum="3ab291282f3dc11dd9e652008ea13851") -package org.apache.hadoop.hive.metastore.annotation;