Index: ql/src/test/results/clientpositive/insert_into4.q.out =================================================================== --- ql/src/test/results/clientpositive/insert_into4.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/insert_into4.q.out (working copy) @@ -214,14 +214,14 @@ Map Operator Tree: TableScan alias: insert_into4a - Statistics: Num rows: 10 Data size: 104 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 20 Data size: 208 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: int), value (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 104 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 20 Data size: 208 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 104 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 20 Data size: 208 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Index: ql/src/test/results/clientpositive/alter_partition_coltype.q.out =================================================================== --- ql/src/test/results/clientpositive/alter_partition_coltype.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/alter_partition_coltype.q.out (working copy) @@ -1007,10 +1007,10 @@ #### A masked pattern was here #### name pt.alterdynamic_part_table numFiles 2 - numRows 1 + numRows 2 partition_columns partcol1/partcol2 partition_columns.types int:string - rawDataSize 2 + rawDataSize 3 serialization.ddl struct alterdynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1039,12 +1039,12 @@ Processor Tree: TableScan alias: alterdynamic_part_table - Statistics: Num rows: 1 Data size: 2 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 3 Basic stats: COMPLETE Column stats: NONE GatherStats: false Select Operator expressions: intcol (type: string) outputColumnNames: _col0 - Statistics: Num rows: 1 Data size: 2 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 2 Data size: 3 Basic stats: COMPLETE Column stats: NONE ListSink PREHOOK: query: explain extended select intcol from pt.alterdynamic_part_table where (partcol1='2' and partcol2='1')or (partcol1='1' and partcol2='__HIVE_DEFAULT_PARTITION__') Index: ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out =================================================================== --- ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out (working copy) @@ -288,8 +288,6 @@ #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -309,8 +307,6 @@ #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Index: ql/src/test/results/clientpositive/insert_into1.q.out =================================================================== --- ql/src/test/results/clientpositive/insert_into1.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/insert_into1.q.out (working copy) @@ -94,6 +94,31 @@ POSTHOOK: Input: default@insert_into1 #### A masked pattern was here #### 10226524244 +PREHOOK: query: explain +select count(*) from insert_into1 +PREHOOK: type: QUERY +POSTHOOK: query: explain +select count(*) from insert_into1 +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: select count(*) from insert_into1 +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +POSTHOOK: query: select count(*) from insert_into1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +100 PREHOOK: query: EXPLAIN INSERT INTO TABLE insert_into1 SELECT * FROM src LIMIT 100 PREHOOK: type: QUERY POSTHOOK: query: EXPLAIN INSERT INTO TABLE insert_into1 SELECT * FROM src LIMIT 100 @@ -178,11 +203,27 @@ POSTHOOK: Input: default@insert_into1 #### A masked pattern was here #### 20453048488 -PREHOOK: query: SELECT COUNT(*) FROM insert_into1 +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: select count(*) from insert_into1 +PREHOOK: type: QUERY PREHOOK: Input: default@insert_into1 #### A masked pattern was here #### -POSTHOOK: query: SELECT COUNT(*) FROM insert_into1 +POSTHOOK: query: select count(*) from insert_into1 POSTHOOK: type: QUERY POSTHOOK: Input: default@insert_into1 #### A masked pattern was here #### @@ -271,6 +312,31 @@ POSTHOOK: Input: default@insert_into1 #### A masked pattern was here #### -826625916 +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 +PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: select count(*) from insert_into1 +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +POSTHOOK: query: select count(*) from insert_into1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +10 PREHOOK: query: DROP TABLE insert_into1 PREHOOK: type: DROPTABLE PREHOOK: Input: default@insert_into1 Index: ql/src/test/results/clientpositive/tez/vectorized_timestamp_funcs.q.out =================================================================== --- ql/src/test/results/clientpositive/tez/vectorized_timestamp_funcs.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/tez/vectorized_timestamp_funcs.q.out (working copy) @@ -545,15 +545,15 @@ Map Operator Tree: TableScan alias: alltypesorc_wrong - Statistics: Num rows: 1 Data size: 103 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 294 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: to_unix_timestamp(stimestamp1) (type: bigint), year(stimestamp1) (type: int), month(stimestamp1) (type: int), day(stimestamp1) (type: int), dayofmonth(stimestamp1) (type: int), weekofyear(stimestamp1) (type: int), hour(stimestamp1) (type: int), minute(stimestamp1) (type: int), second(stimestamp1) (type: int) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 103 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 294 Basic stats: COMPLETE Column stats: NONE Reduce Output Operator key expressions: _col0 (type: bigint) sort order: + - Statistics: Num rows: 1 Data size: 103 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 294 Basic stats: COMPLETE Column stats: NONE value expressions: _col1 (type: int), _col2 (type: int), _col3 (type: int), _col4 (type: int), _col5 (type: int), _col6 (type: int), _col7 (type: int), _col8 (type: int) Execution mode: vectorized Reducer 2 @@ -561,10 +561,10 @@ Select Operator expressions: KEY.reducesinkkey0 (type: bigint), VALUE._col0 (type: int), VALUE._col1 (type: int), VALUE._col2 (type: int), VALUE._col3 (type: int), VALUE._col4 (type: int), VALUE._col5 (type: int), VALUE._col6 (type: int), VALUE._col7 (type: int) outputColumnNames: _col0, _col1, _col2, _col3, _col4, _col5, _col6, _col7, _col8 - Statistics: Num rows: 1 Data size: 103 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 294 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 1 Data size: 103 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 3 Data size: 294 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Index: ql/src/test/results/clientpositive/tez/insert_into1.q.out =================================================================== --- ql/src/test/results/clientpositive/tez/insert_into1.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/tez/insert_into1.q.out (working copy) @@ -104,6 +104,31 @@ POSTHOOK: Input: default@insert_into1 #### A masked pattern was here #### 10226524244 +PREHOOK: query: explain +select count(*) from insert_into1 +PREHOOK: type: QUERY +POSTHOOK: query: explain +select count(*) from insert_into1 +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: select count(*) from insert_into1 +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +POSTHOOK: query: select count(*) from insert_into1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +100 PREHOOK: query: EXPLAIN INSERT INTO TABLE insert_into1 SELECT * FROM src LIMIT 100 PREHOOK: type: QUERY POSTHOOK: query: EXPLAIN INSERT INTO TABLE insert_into1 SELECT * FROM src LIMIT 100 @@ -198,11 +223,27 @@ POSTHOOK: Input: default@insert_into1 #### A masked pattern was here #### 20453048488 -PREHOOK: query: SELECT COUNT(*) FROM insert_into1 +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: select count(*) from insert_into1 +PREHOOK: type: QUERY PREHOOK: Input: default@insert_into1 #### A masked pattern was here #### -POSTHOOK: query: SELECT COUNT(*) FROM insert_into1 +POSTHOOK: query: select count(*) from insert_into1 POSTHOOK: type: QUERY POSTHOOK: Input: default@insert_into1 #### A masked pattern was here #### @@ -301,6 +342,31 @@ POSTHOOK: Input: default@insert_into1 #### A masked pattern was here #### -826625916 +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 +PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into1 +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: select count(*) from insert_into1 +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +POSTHOOK: query: select count(*) from insert_into1 +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into1 +#### A masked pattern was here #### +10 PREHOOK: query: DROP TABLE insert_into1 PREHOOK: type: DROPTABLE PREHOOK: Input: default@insert_into1 Index: ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out =================================================================== --- ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/tez/dynpart_sort_opt_vectorization.q.out (working copy) @@ -930,8 +930,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 16 - rawDataSize 320 + numRows 32 + rawDataSize 640 totalSize 1348 #### A masked pattern was here #### @@ -974,8 +974,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 3 - rawDataSize 60 + numRows 6 + rawDataSize 120 totalSize 1050 #### A masked pattern was here #### @@ -1018,8 +1018,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 7 - rawDataSize 140 + numRows 14 + rawDataSize 280 totalSize 1166 #### A masked pattern was here #### @@ -1062,8 +1062,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 3 - rawDataSize 60 + numRows 6 + rawDataSize 120 totalSize 1050 #### A masked pattern was here #### @@ -1105,8 +1105,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 16 - rawDataSize 320 + numRows 32 + rawDataSize 640 totalSize 4340 #### A masked pattern was here #### @@ -1148,8 +1148,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 3 - rawDataSize 60 + numRows 6 + rawDataSize 120 totalSize 2094 #### A masked pattern was here #### @@ -1191,8 +1191,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 16 - rawDataSize 320 + numRows 32 + rawDataSize 640 totalSize 4326 #### A masked pattern was here #### @@ -1234,8 +1234,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 3 - rawDataSize 60 + numRows 6 + rawDataSize 120 totalSize 2094 #### A masked pattern was here #### Index: ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out =================================================================== --- ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/tez/dynpart_sort_optimization.q.out (working copy) @@ -843,8 +843,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 16 - rawDataSize 415 + numRows 32 + rawDataSize 830 totalSize 862 #### A masked pattern was here #### @@ -887,8 +887,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 3 - rawDataSize 78 + numRows 6 + rawDataSize 156 totalSize 162 #### A masked pattern was here #### @@ -931,8 +931,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 7 - rawDataSize 181 + numRows 14 + rawDataSize 362 totalSize 376 #### A masked pattern was here #### @@ -975,8 +975,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 3 - rawDataSize 78 + numRows 6 + rawDataSize 156 totalSize 162 #### A masked pattern was here #### @@ -1018,8 +1018,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 16 - rawDataSize 415 + numRows 32 + rawDataSize 830 totalSize 862 #### A masked pattern was here #### @@ -1061,8 +1061,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 3 - rawDataSize 78 + numRows 6 + rawDataSize 156 totalSize 162 #### A masked pattern was here #### @@ -1104,8 +1104,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 16 - rawDataSize 415 + numRows 32 + rawDataSize 830 totalSize 862 #### A masked pattern was here #### @@ -1147,8 +1147,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 3 - rawDataSize 78 + numRows 6 + rawDataSize 156 totalSize 162 #### A masked pattern was here #### Index: ql/src/test/results/clientpositive/tez/auto_sortmerge_join_5.q.out =================================================================== --- ql/src/test/results/clientpositive/tez/auto_sortmerge_join_5.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/tez/auto_sortmerge_join_5.q.out (working copy) @@ -170,8 +170,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -192,8 +190,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -240,8 +236,6 @@ #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -262,8 +256,6 @@ #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -409,8 +401,6 @@ #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -431,8 +421,6 @@ #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -499,8 +487,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -521,8 +507,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -668,8 +652,6 @@ #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -690,8 +672,6 @@ #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -758,8 +738,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -780,8 +758,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Index: ql/src/test/results/clientpositive/tez/insert_into2.q.out =================================================================== --- ql/src/test/results/clientpositive/tez/insert_into2.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/tez/insert_into2.q.out (working copy) @@ -97,6 +97,31 @@ POSTHOOK: Output: default@insert_into2@ds=1 POSTHOOK: Lineage: insert_into2 PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: insert_into2 PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain +select count (*) from insert_into2 where ds = '1' +PREHOOK: type: QUERY +POSTHOOK: query: explain +select count (*) from insert_into2 where ds = '1' +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: select count (*) from insert_into2 where ds = '1' +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into2 +#### A masked pattern was here #### +POSTHOOK: query: select count (*) from insert_into2 where ds = '1' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into2 +#### A masked pattern was here #### +100 PREHOOK: query: INSERT INTO TABLE insert_into2 PARTITION (ds='1') SELECT * FROM src limit 100 PREHOOK: type: QUERY PREHOOK: Input: default@src @@ -107,15 +132,29 @@ POSTHOOK: Output: default@insert_into2@ds=1 POSTHOOK: Lineage: insert_into2 PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: insert_into2 PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='1' +PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='1' +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + PREHOOK: query: SELECT COUNT(*) FROM insert_into2 WHERE ds='1' PREHOOK: type: QUERY PREHOOK: Input: default@insert_into2 -PREHOOK: Input: default@insert_into2@ds=1 #### A masked pattern was here #### POSTHOOK: query: SELECT COUNT(*) FROM insert_into2 WHERE ds='1' POSTHOOK: type: QUERY POSTHOOK: Input: default@insert_into2 -POSTHOOK: Input: default@insert_into2@ds=1 #### A masked pattern was here #### 200 PREHOOK: query: SELECT SUM(HASH(c)) FROM ( @@ -237,6 +276,31 @@ POSTHOOK: Input: default@insert_into2@ds=2 #### A masked pattern was here #### -36239931656 +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into2 +#### A masked pattern was here #### +100 PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE insert_into2 PARTITION (ds='2') SELECT * FROM src LIMIT 50 PREHOOK: type: QUERY @@ -341,6 +405,31 @@ POSTHOOK: Input: default@insert_into2@ds=2 #### A masked pattern was here #### -27100860056 +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into2 +#### A masked pattern was here #### +50 PREHOOK: query: DROP TABLE insert_into2 PREHOOK: type: DROPTABLE PREHOOK: Input: default@insert_into2 Index: ql/src/test/results/clientpositive/tez/alter_merge_stats_orc.q.out =================================================================== --- ql/src/test/results/clientpositive/tez/alter_merge_stats_orc.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/tez/alter_merge_stats_orc.q.out (working copy) @@ -247,8 +247,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 3 - numRows 500 - rawDataSize 47000 + numRows 1500 + rawDataSize 141000 totalSize 7488 #### A masked pattern was here #### Index: ql/src/test/results/clientpositive/mapjoin_test_outer.q.out =================================================================== --- ql/src/test/results/clientpositive/mapjoin_test_outer.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/mapjoin_test_outer.q.out (working copy) @@ -275,7 +275,7 @@ src2 TableScan alias: src2 - Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 9 Data size: 40 Basic stats: COMPLETE Column stats: NONE HashTable Sink Operator condition expressions: 0 {value} @@ -1108,7 +1108,7 @@ src2 TableScan alias: src2 - Statistics: Num rows: 1 Data size: 13 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 9 Data size: 40 Basic stats: COMPLETE Column stats: NONE HashTable Sink Operator condition expressions: 0 {value} Index: ql/src/test/results/clientpositive/bucketmapjoin4.q.out =================================================================== --- ql/src/test/results/clientpositive/bucketmapjoin4.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/bucketmapjoin4.q.out (working copy) @@ -292,8 +292,6 @@ #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -313,8 +311,6 @@ #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -783,8 +779,6 @@ #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -804,8 +798,6 @@ #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Index: ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out =================================================================== --- ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out (working copy) @@ -302,8 +302,6 @@ #### A masked pattern was here #### name default.test1 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -324,8 +322,6 @@ #### A masked pattern was here #### name default.test1 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -501,8 +497,6 @@ #### A masked pattern was here #### name default.test2 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -523,8 +517,6 @@ #### A masked pattern was here #### name default.test2 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -689,8 +681,6 @@ #### A masked pattern was here #### name default.test1 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -711,8 +701,6 @@ #### A masked pattern was here #### name default.test1 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -880,8 +868,6 @@ #### A masked pattern was here #### name default.test1 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -902,8 +888,6 @@ #### A masked pattern was here #### name default.test1 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1071,8 +1055,6 @@ #### A masked pattern was here #### name default.test1 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1093,8 +1075,6 @@ #### A masked pattern was here #### name default.test1 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1262,8 +1242,6 @@ #### A masked pattern was here #### name default.test1 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1284,8 +1262,6 @@ #### A masked pattern was here #### name default.test1 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1453,8 +1429,6 @@ #### A masked pattern was here #### name default.test2 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1475,8 +1449,6 @@ #### A masked pattern was here #### name default.test2 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1644,8 +1616,6 @@ #### A masked pattern was here #### name default.test2 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1666,8 +1636,6 @@ #### A masked pattern was here #### name default.test2 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1835,8 +1803,6 @@ #### A masked pattern was here #### name default.test3 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -1857,8 +1823,6 @@ #### A masked pattern was here #### name default.test3 numFiles 3 - numRows 0 - rawDataSize 0 serialization.ddl struct test3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Index: ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out =================================================================== --- ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/dynpart_sort_optimization.q.out (working copy) @@ -783,8 +783,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 16 - rawDataSize 415 + numRows 32 + rawDataSize 830 totalSize 862 #### A masked pattern was here #### @@ -827,8 +827,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 3 - rawDataSize 78 + numRows 6 + rawDataSize 156 totalSize 162 #### A masked pattern was here #### @@ -871,8 +871,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 7 - rawDataSize 181 + numRows 14 + rawDataSize 362 totalSize 376 #### A masked pattern was here #### @@ -915,8 +915,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 3 - rawDataSize 78 + numRows 6 + rawDataSize 156 totalSize 162 #### A masked pattern was here #### @@ -958,8 +958,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 16 - rawDataSize 415 + numRows 32 + rawDataSize 830 totalSize 862 #### A masked pattern was here #### @@ -1001,8 +1001,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 3 - rawDataSize 78 + numRows 6 + rawDataSize 156 totalSize 162 #### A masked pattern was here #### @@ -1044,8 +1044,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 16 - rawDataSize 415 + numRows 32 + rawDataSize 830 totalSize 862 #### A masked pattern was here #### @@ -1087,8 +1087,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 3 - rawDataSize 78 + numRows 6 + rawDataSize 156 totalSize 162 #### A masked pattern was here #### Index: ql/src/test/results/clientpositive/insert_into5.q.out =================================================================== --- ql/src/test/results/clientpositive/insert_into5.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/insert_into5.q.out (working copy) @@ -233,14 +233,14 @@ Map Operator Tree: TableScan alias: insert_into5a - Statistics: Num rows: 10 Data size: 50 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 20 Data size: 100 Basic stats: COMPLETE Column stats: NONE Select Operator expressions: key (type: int), value (type: string) outputColumnNames: _col0, _col1 - Statistics: Num rows: 10 Data size: 50 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 20 Data size: 100 Basic stats: COMPLETE Column stats: NONE File Output Operator compressed: false - Statistics: Num rows: 10 Data size: 50 Basic stats: COMPLETE Column stats: NONE + Statistics: Num rows: 20 Data size: 100 Basic stats: COMPLETE Column stats: NONE table: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat Index: ql/src/test/results/clientpositive/columnstats_tbllvl.q.out =================================================================== --- ql/src/test/results/clientpositive/columnstats_tbllvl.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/columnstats_tbllvl.q.out (working copy) @@ -146,8 +146,6 @@ #### A masked pattern was here #### name default.uservisits_web_text_none numFiles 1 - numRows 0 - rawDataSize 0 serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -167,8 +165,6 @@ #### A masked pattern was here #### name default.uservisits_web_text_none numFiles 1 - numRows 0 - rawDataSize 0 serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Index: ql/src/test/results/clientpositive/bucketmapjoin1.q.out =================================================================== --- ql/src/test/results/clientpositive/bucketmapjoin1.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/bucketmapjoin1.q.out (working copy) @@ -601,8 +601,6 @@ #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -622,8 +620,6 @@ #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Index: ql/src/test/results/clientpositive/auto_sortmerge_join_5.q.out =================================================================== --- ql/src/test/results/clientpositive/auto_sortmerge_join_5.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/auto_sortmerge_join_5.q.out (working copy) @@ -157,8 +157,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -179,8 +177,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -329,8 +325,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -351,8 +345,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -532,8 +524,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -554,8 +544,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -594,8 +582,6 @@ #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -709,8 +695,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -731,8 +715,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -771,8 +753,6 @@ #### A masked pattern was here #### name default.bucket_small numFiles 4 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -863,8 +843,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -885,8 +863,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Index: ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out =================================================================== --- ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/display_colstats_tbllvl.q.out (working copy) @@ -162,8 +162,6 @@ #### A masked pattern was here #### name default.uservisits_web_text_none numFiles 1 - numRows 0 - rawDataSize 0 serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -183,8 +181,6 @@ #### A masked pattern was here #### name default.uservisits_web_text_none numFiles 1 - numRows 0 - rawDataSize 0 serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Index: ql/src/test/results/clientpositive/insert_into2.q.out =================================================================== --- ql/src/test/results/clientpositive/insert_into2.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/insert_into2.q.out (working copy) @@ -87,6 +87,31 @@ POSTHOOK: Output: default@insert_into2@ds=1 POSTHOOK: Lineage: insert_into2 PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: insert_into2 PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain +select count (*) from insert_into2 where ds = '1' +PREHOOK: type: QUERY +POSTHOOK: query: explain +select count (*) from insert_into2 where ds = '1' +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: select count (*) from insert_into2 where ds = '1' +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into2 +#### A masked pattern was here #### +POSTHOOK: query: select count (*) from insert_into2 where ds = '1' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into2 +#### A masked pattern was here #### +100 PREHOOK: query: INSERT INTO TABLE insert_into2 PARTITION (ds='1') SELECT * FROM src limit 100 PREHOOK: type: QUERY PREHOOK: Input: default@src @@ -97,15 +122,29 @@ POSTHOOK: Output: default@insert_into2@ds=1 POSTHOOK: Lineage: insert_into2 PARTITION(ds=1).key EXPRESSION [(src)src.FieldSchema(name:key, type:string, comment:default), ] POSTHOOK: Lineage: insert_into2 PARTITION(ds=1).value SIMPLE [(src)src.FieldSchema(name:value, type:string, comment:default), ] +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='1' +PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='1' +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + PREHOOK: query: SELECT COUNT(*) FROM insert_into2 WHERE ds='1' PREHOOK: type: QUERY PREHOOK: Input: default@insert_into2 -PREHOOK: Input: default@insert_into2@ds=1 #### A masked pattern was here #### POSTHOOK: query: SELECT COUNT(*) FROM insert_into2 WHERE ds='1' POSTHOOK: type: QUERY POSTHOOK: Input: default@insert_into2 -POSTHOOK: Input: default@insert_into2@ds=1 #### A masked pattern was here #### 200 PREHOOK: query: SELECT SUM(HASH(c)) FROM ( @@ -217,6 +256,31 @@ POSTHOOK: Input: default@insert_into2@ds=2 #### A masked pattern was here #### -36239931656 +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into2 +#### A masked pattern was here #### +100 PREHOOK: query: EXPLAIN INSERT OVERWRITE TABLE insert_into2 PARTITION (ds='2') SELECT * FROM src LIMIT 50 PREHOOK: type: QUERY @@ -311,6 +375,31 @@ POSTHOOK: Input: default@insert_into2@ds=2 #### A masked pattern was here #### -27100860056 +PREHOOK: query: explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +PREHOOK: type: QUERY +POSTHOOK: query: explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +POSTHOOK: type: QUERY +STAGE DEPENDENCIES: + Stage-0 is a root stage + +STAGE PLANS: + Stage: Stage-0 + Fetch Operator + limit: 1 + Processor Tree: + ListSink + +PREHOOK: query: SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +PREHOOK: type: QUERY +PREHOOK: Input: default@insert_into2 +#### A masked pattern was here #### +POSTHOOK: query: SELECT COUNT(*) FROM insert_into2 WHERE ds='2' +POSTHOOK: type: QUERY +POSTHOOK: Input: default@insert_into2 +#### A masked pattern was here #### +50 PREHOOK: query: DROP TABLE insert_into2 PREHOOK: type: DROPTABLE PREHOOK: Input: default@insert_into2 Index: ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out =================================================================== --- ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out (working copy) @@ -343,8 +343,6 @@ #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -364,8 +362,6 @@ #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Index: ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out =================================================================== --- ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/dynpart_sort_opt_vectorization.q.out (working copy) @@ -878,8 +878,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 16 - rawDataSize 320 + numRows 32 + rawDataSize 640 totalSize 1348 #### A masked pattern was here #### @@ -922,8 +922,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 3 - rawDataSize 60 + numRows 6 + rawDataSize 120 totalSize 1050 #### A masked pattern was here #### @@ -966,8 +966,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 7 - rawDataSize 140 + numRows 14 + rawDataSize 280 totalSize 1166 #### A masked pattern was here #### @@ -1010,8 +1010,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 2 - numRows 3 - rawDataSize 60 + numRows 6 + rawDataSize 120 totalSize 1050 #### A masked pattern was here #### @@ -1053,8 +1053,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 16 - rawDataSize 320 + numRows 32 + rawDataSize 640 totalSize 4340 #### A masked pattern was here #### @@ -1096,8 +1096,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 3 - rawDataSize 60 + numRows 6 + rawDataSize 120 totalSize 2094 #### A masked pattern was here #### @@ -1139,8 +1139,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 16 - rawDataSize 320 + numRows 32 + rawDataSize 640 totalSize 4326 #### A masked pattern was here #### @@ -1182,8 +1182,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 8 - numRows 3 - rawDataSize 60 + numRows 6 + rawDataSize 120 totalSize 2094 #### A masked pattern was here #### Index: ql/src/test/results/clientpositive/bucketcontext_5.q.out =================================================================== --- ql/src/test/results/clientpositive/bucketcontext_5.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/bucketcontext_5.q.out (working copy) @@ -200,8 +200,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -222,8 +220,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -380,8 +376,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -402,8 +396,6 @@ #### A masked pattern was here #### name default.bucket_big numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Index: ql/src/test/results/clientpositive/stats11.q.out =================================================================== --- ql/src/test/results/clientpositive/stats11.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/stats11.q.out (working copy) @@ -535,8 +535,6 @@ #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe @@ -556,8 +554,6 @@ #### A masked pattern was here #### name default.srcbucket_mapjoin numFiles 2 - numRows 0 - rawDataSize 0 serialization.ddl struct srcbucket_mapjoin { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe Index: ql/src/test/results/clientpositive/alter_merge_stats_orc.q.out =================================================================== --- ql/src/test/results/clientpositive/alter_merge_stats_orc.q.out (revision 1628879) +++ ql/src/test/results/clientpositive/alter_merge_stats_orc.q.out (working copy) @@ -247,8 +247,8 @@ Partition Parameters: COLUMN_STATS_ACCURATE true numFiles 3 - numRows 500 - rawDataSize 47000 + numRows 1500 + rawDataSize 141000 totalSize 7488 #### A masked pattern was here #### Index: ql/src/test/queries/clientpositive/insert_into1.q =================================================================== --- ql/src/test/queries/clientpositive/insert_into1.q (revision 1628879) +++ ql/src/test/queries/clientpositive/insert_into1.q (working copy) @@ -1,3 +1,4 @@ +set hive.compute.query.using.stats=true; DROP TABLE insert_into1; CREATE TABLE insert_into1 (key int, value string); @@ -7,14 +8,18 @@ SELECT SUM(HASH(c)) FROM ( SELECT TRANSFORM(*) USING 'tr \t _' AS (c) FROM insert_into1 ) t; - +explain +select count(*) from insert_into1; +select count(*) from insert_into1; EXPLAIN INSERT INTO TABLE insert_into1 SELECT * FROM src LIMIT 100; INSERT INTO TABLE insert_into1 SELECT * FROM src LIMIT 100; SELECT SUM(HASH(c)) FROM ( SELECT TRANSFORM(*) USING 'tr \t _' AS (c) FROM insert_into1 ) t; +explain SELECT COUNT(*) FROM insert_into1; +select count(*) from insert_into1; EXPLAIN INSERT OVERWRITE TABLE insert_into1 SELECT * FROM src LIMIT 10; INSERT OVERWRITE TABLE insert_into1 SELECT * FROM src LIMIT 10; @@ -22,5 +27,10 @@ SELECT TRANSFORM(*) USING 'tr \t _' AS (c) FROM insert_into1 ) t; +explain +SELECT COUNT(*) FROM insert_into1; +select count(*) from insert_into1; DROP TABLE insert_into1; + +set hive.compute.query.using.stats=false; Index: ql/src/test/queries/clientpositive/insert_into2.q =================================================================== --- ql/src/test/queries/clientpositive/insert_into2.q (revision 1628879) +++ ql/src/test/queries/clientpositive/insert_into2.q (working copy) @@ -1,3 +1,4 @@ +set hive.compute.query.using.stats=true; DROP TABLE insert_into2; CREATE TABLE insert_into2 (key int, value string) PARTITIONED BY (ds string); @@ -5,8 +6,13 @@ EXPLAIN INSERT INTO TABLE insert_into2 PARTITION (ds='1') SELECT * FROM src LIMIT 100; INSERT INTO TABLE insert_into2 PARTITION (ds='1') SELECT * FROM src limit 100; +explain +select count (*) from insert_into2 where ds = '1'; +select count (*) from insert_into2 where ds = '1'; INSERT INTO TABLE insert_into2 PARTITION (ds='1') SELECT * FROM src limit 100; +explain SELECT COUNT(*) FROM insert_into2 WHERE ds='1'; +SELECT COUNT(*) FROM insert_into2 WHERE ds='1'; SELECT SUM(HASH(c)) FROM ( SELECT TRANSFORM(*) USING 'tr \t _' AS (c) FROM insert_into2 ) t; @@ -19,6 +25,9 @@ SELECT SUM(HASH(c)) FROM ( SELECT TRANSFORM(*) USING 'tr \t _' AS (c) FROM insert_into2 ) t; +explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='2'; +SELECT COUNT(*) FROM insert_into2 WHERE ds='2'; EXPLAIN INSERT OVERWRITE TABLE insert_into2 PARTITION (ds='2') SELECT * FROM src LIMIT 50; @@ -27,5 +36,11 @@ SELECT SUM(HASH(c)) FROM ( SELECT TRANSFORM(*) USING 'tr \t _' AS (c) FROM insert_into2 ) t; +explain +SELECT COUNT(*) FROM insert_into2 WHERE ds='2'; +SELECT COUNT(*) FROM insert_into2 WHERE ds='2'; + DROP TABLE insert_into2; + +set hive.compute.query.using.stats=false; Index: ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java =================================================================== --- ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java (revision 1628879) +++ ql/src/java/org/apache/hadoop/hive/ql/metadata/Hive.java (working copy) @@ -1565,6 +1565,7 @@ tbl.replaceFiles(loadPath, isSrcLocal); } else { tbl.copyFiles(loadPath, isSrcLocal, isAcid); + tbl.getParameters().put(StatsSetupConst.STATS_GENERATED_VIA_STATS_TASK, "true"); } try { @@ -1678,17 +1679,6 @@ return getPartition(tbl, partSpec, forceCreate, null, true); } - private static void clearPartitionStats(org.apache.hadoop.hive.metastore.api.Partition tpart) { - Map tpartParams = tpart.getParameters(); - if (tpartParams == null) { - return; - } - - for (String statType : StatsSetupConst.supportedStats) { - tpartParams.remove(statType); - } - } - /** * Returns partition metadata * @@ -1756,7 +1746,7 @@ throw new HiveException("new partition path should not be null or empty."); } tpart.getSd().setLocation(partPath); - clearPartitionStats(tpart); + tpart.getParameters().put(StatsSetupConst.STATS_GENERATED_VIA_STATS_TASK,"true"); String fullName = tbl.getTableName(); if (!org.apache.commons.lang.StringUtils.isEmpty(tbl.getDbName())) { fullName = tbl.getDbName() + "." + tbl.getTableName();