From 0b5f26b248c6bbfe60b24a34cf1b1c90e29eb6e1 Mon Sep 17 00:00:00 2001 From: Ashutosh Chauhan Date: Fri, 15 May 2020 23:02:35 -0700 Subject: [PATCH] HIVE-23292 : Reduce PartitionDesc payload in MapWork --- .../insert_into_dynamic_partitions.q.out | 8 - .../clientpositive/insert_into_table.q.out | 76 --- .../insert_overwrite_directory.q.out | 32 - .../insert_overwrite_dynamic_partitions.q.out | 8 - .../insert_overwrite_table.q.out | 76 --- .../write_final_output_blobstore.q.out | 64 -- .../hadoop/hive/ql/plan/PartitionDesc.java | 7 +- .../apache/hadoop/hive/ql/plan/TableDesc.java | 2 - .../beeline/smb_mapjoin_11.q.out | 34 - .../beeline/smb_mapjoin_12.q.out | 28 - .../beeline/smb_mapjoin_13.q.out | 30 - .../clientpositive/binary_output_format.q.out | 80 --- .../clientpositive/bucket_map_join_1.q.out | 13 - .../clientpositive/bucket_map_join_2.q.out | 13 - .../bucket_map_join_spark1.q.out | 86 --- .../bucket_map_join_spark2.q.out | 86 --- .../bucket_map_join_spark3.q.out | 86 --- .../bucket_map_join_spark4.q.out | 90 --- .../clientpositive/bucketcontext_1.q.out | 45 -- .../clientpositive/bucketcontext_2.q.out | 45 -- .../clientpositive/bucketcontext_3.q.out | 36 -- .../clientpositive/bucketcontext_4.q.out | 36 -- .../clientpositive/bucketcontext_5.q.out | 26 - .../clientpositive/bucketcontext_6.q.out | 36 -- .../clientpositive/bucketcontext_7.q.out | 54 -- .../clientpositive/bucketcontext_8.q.out | 54 -- .../clientpositive/bucketmapjoin10.q.out | 36 -- .../clientpositive/bucketmapjoin11.q.out | 72 --- .../clientpositive/bucketmapjoin12.q.out | 38 -- .../clientpositive/bucketmapjoin13.q.out | 90 --- .../clientpositive/bucketmapjoin5.q.out | 164 ----- .../clientpositive/bucketmapjoin8.q.out | 36 -- .../clientpositive/bucketmapjoin9.q.out | 36 -- .../bucketmapjoin_negative.q.out | 86 --- .../bucketmapjoin_negative2.q.out | 95 --- .../bucketmapjoin_negative3.q.out | 117 ---- .../druid/druidkafkamini_basic.q.out | 32 - .../druid/druidmini_expressions.q.out | 16 - .../encryption_join_unencrypted_tbl.q.out | 32 - ..._join_with_different_encryption_keys.q.out | 32 - .../erasurecoding/erasure_explain.q.out | 48 -- .../results/clientpositive/join_map_ppr.q.out | 152 ----- .../kafka/kafka_storage_handler.q.out | 28 - .../clientpositive/list_bucket_dml_8.q.out | 53 -- .../llap/acid_bucket_pruning.q.out | 15 - .../clientpositive/llap/acid_nullscan.q.out | 15 - .../llap/alter_partition_coltype.q.out | 48 -- .../llap/analyze_table_null_partition.q.out | 48 -- .../llap/autoColumnStats_5a.q.out | 20 - .../llap/autoColumnStats_8.q.out | 60 -- .../llap/auto_join_reordering_values.q.out | 80 --- .../llap/auto_sortmerge_join_1.q.out | 81 --- .../llap/auto_sortmerge_join_11.q.out | 126 ---- .../llap/auto_sortmerge_join_12.q.out | 45 -- .../llap/auto_sortmerge_join_2.q.out | 54 -- .../llap/auto_sortmerge_join_3.q.out | 81 --- .../llap/auto_sortmerge_join_4.q.out | 81 --- .../llap/auto_sortmerge_join_5.q.out | 78 --- .../llap/auto_sortmerge_join_7.q.out | 108 ---- .../llap/auto_sortmerge_join_8.q.out | 108 ---- .../results/clientpositive/llap/bucket1.q.out | 30 - .../results/clientpositive/llap/bucket2.q.out | 30 - .../results/clientpositive/llap/bucket3.q.out | 20 - .../results/clientpositive/llap/bucket4.q.out | 30 - .../results/clientpositive/llap/bucket5.q.out | 94 --- .../clientpositive/llap/bucket_many.q.out | 30 - .../llap/bucket_map_join_tez2.q.out | 16 - .../llap/bucket_num_reducers.q.out | 30 - .../llap/bucket_num_reducers2.q.out | 30 - .../clientpositive/llap/bucketmapjoin1.q.out | 76 --- .../clientpositive/llap/bucketmapjoin2.q.out | 111 ---- .../clientpositive/llap/bucketmapjoin3.q.out | 68 -- .../clientpositive/llap/bucketmapjoin4.q.out | 84 --- .../clientpositive/llap/bucketmapjoin7.q.out | 18 - .../llap/cbo_rp_outer_join_ppr.q.out | 80 --- .../llap/cbo_stats_estimation.q.out | 28 - .../llap/column_table_stats.q.out | 69 -- .../llap/column_table_stats_orc.q.out | 49 -- .../llap/columnstats_partlvl.q.out | 22 - .../llap/columnstats_tbllvl.q.out | 28 - .../clientpositive/llap/comments.q.out | 32 - .../llap/constantPropagateForSubQuery.q.out | 32 - .../llap/disable_merge_for_bucketing.q.out | 30 - .../llap/display_colstats_tbllvl.q.out | 14 - .../llap/dynamic_partition_skip_default.q.out | 48 -- .../llap/dynamic_semijoin_reduction.q.out | 44 -- .../llap/dynamic_semijoin_user_level.q.out | 44 -- .../llap/extrapolate_part_stats_date.q.out | 48 -- .../llap/extrapolate_part_stats_full.q.out | 144 ----- .../llap/extrapolate_part_stats_partial.q.out | 456 -------------- .../extrapolate_part_stats_partial_ndv.q.out | 228 ------- .../clientpositive/llap/filter_aggr.q.out | 16 - .../llap/filter_join_breaktask.q.out | 36 -- .../clientpositive/llap/filter_union.q.out | 32 - .../clientpositive/llap/groupby_map_ppr.q.out | 40 -- .../llap/groupby_map_ppr_multi_distinct.q.out | 40 -- .../clientpositive/llap/groupby_ppr.q.out | 40 -- .../llap/groupby_ppr_multi_distinct.q.out | 80 --- .../llap/groupby_sort_1_23.q.out | 587 ------------------ .../clientpositive/llap/groupby_sort_6.q.out | 59 -- .../llap/groupby_sort_skew_1_23.q.out | 587 ------------------ .../llap/infer_bucket_sort_num_buckets.q.out | 30 - .../results/clientpositive/llap/input23.q.out | 12 - .../results/clientpositive/llap/input42.q.out | 72 --- .../clientpositive/llap/input_part1.q.out | 28 - .../clientpositive/llap/input_part2.q.out | 56 -- .../clientpositive/llap/input_part7.q.out | 48 -- .../clientpositive/llap/input_part9.q.out | 24 - .../results/clientpositive/llap/join17.q.out | 48 -- .../results/clientpositive/llap/join26.q.out | 60 -- .../results/clientpositive/llap/join32.q.out | 60 -- .../clientpositive/llap/join32_lessSize.q.out | 260 -------- .../results/clientpositive/llap/join33.q.out | 60 -- .../results/clientpositive/llap/join34.q.out | 72 --- .../results/clientpositive/llap/join35.q.out | 72 --- .../results/clientpositive/llap/join9.q.out | 44 -- .../llap/join_filters_overlap.q.out | 160 ----- .../llap/list_bucket_dml_1.q.out | 42 -- .../llap/list_bucket_dml_10.q.out | 22 - .../llap/list_bucket_dml_11.q.out | 34 - .../llap/list_bucket_dml_12.q.out | 46 -- .../llap/list_bucket_dml_13.q.out | 34 - .../llap/list_bucket_dml_14.q.out | 32 - .../llap/list_bucket_dml_2.q.out | 42 -- .../llap/list_bucket_dml_3.q.out | 42 -- .../llap/list_bucket_dml_4.q.out | 72 --- .../llap/list_bucket_dml_5.q.out | 54 -- .../llap/list_bucket_dml_6.q.out | 84 --- .../llap/list_bucket_dml_7.q.out | 84 --- .../llap/list_bucket_dml_9.q.out | 72 --- .../llap/list_bucket_query_multiskew_1.q.out | 48 -- .../llap/list_bucket_query_multiskew_2.q.out | 36 -- .../llap/list_bucket_query_multiskew_3.q.out | 36 -- .../llap/list_bucket_query_oneskew_1.q.out | 27 - .../llap/list_bucket_query_oneskew_2.q.out | 36 -- .../llap/list_bucket_query_oneskew_3.q.out | 9 - .../clientpositive/llap/llap_nullscan.q.out | 16 - .../clientpositive/llap/load_dyn_part8.q.out | 60 -- .../clientpositive/llap/louter_join_ppr.q.out | 160 ----- .../clientpositive/llap/mapjoin_mapjoin.q.out | 80 --- .../results/clientpositive/llap/merge3.q.out | 76 --- .../clientpositive/llap/metadataonly1.q.out | 240 ------- .../llap/murmur_hash_migration.q.out | 68 -- .../llap/murmur_hash_migration2.q.out | 8 - .../llap/offset_limit_global_optimizer.q.out | 384 ------------ .../llap/optimize_nullscan.q.out | 272 -------- .../clientpositive/llap/outer_join_ppr.q.out | 80 --- .../llap/parquet_vectorization_0.q.out | 80 --- .../llap/partition_pruning.q.out | 60 -- .../results/clientpositive/llap/pcr.q.out | 568 ----------------- .../results/clientpositive/llap/pcs.q.out | 252 -------- .../clientpositive/llap/pointlookup2.q.out | 328 ---------- .../clientpositive/llap/pointlookup3.q.out | 288 --------- .../clientpositive/llap/pointlookup4.q.out | 48 -- .../clientpositive/llap/ppd_join_filter.q.out | 64 -- .../clientpositive/llap/ppd_union_view.q.out | 36 -- .../results/clientpositive/llap/ppd_vc.q.out | 112 ---- .../llap/ppr_allchildsarenull.q.out | 72 --- .../results/clientpositive/llap/push_or.q.out | 24 - .../llap/rand_partitionpruner2.q.out | 40 -- .../llap/rand_partitionpruner3.q.out | 24 - .../llap/reduce_deduplicate.q.out | 36 -- .../clientpositive/llap/router_join_ppr.q.out | 160 ----- .../results/clientpositive/llap/sample1.q.out | 28 - .../clientpositive/llap/sample10.q.out | 40 -- .../results/clientpositive/llap/sample5.q.out | 31 - .../results/clientpositive/llap/sample6.q.out | 136 ---- .../results/clientpositive/llap/sample7.q.out | 31 - .../results/clientpositive/llap/sample8.q.out | 60 -- .../clientpositive/llap/sharedwork.q.out | 112 ---- .../clientpositive/llap/smb_mapjoin_15.q.out | 120 ---- .../results/clientpositive/llap/stats0.q.out | 64 -- .../results/clientpositive/llap/stats11.q.out | 76 --- .../results/clientpositive/llap/stats12.q.out | 18 - .../results/clientpositive/llap/stats13.q.out | 9 - .../results/clientpositive/llap/stats3.q.out | 8 - .../temp_table_alter_partition_coltype.q.out | 120 ---- .../temp_table_display_colstats_tbllvl.q.out | 12 - .../llap/tez_fixed_bucket_pruning.q.out | 156 ----- .../llap/tez_join_result_complex.q.out | 56 -- .../llap/topnkey_windowing.q.out | 32 - .../clientpositive/llap/unionDistinct_1.q.out | 234 ------- .../clientpositive/llap/vectorization_0.q.out | 80 --- .../clientpositive/regexp_extract.q.out | 32 - .../serde_user_properties.q.out | 48 -- .../sort_merge_join_desc_5.q.out | 10 - .../sort_merge_join_desc_6.q.out | 20 - .../sort_merge_join_desc_7.q.out | 40 -- .../temp_table_partition_pruning.q.out | 90 --- .../results/clientpositive/timestamp.q.out | 16 - .../clientpositive/transform_ppr1.q.out | 48 -- .../clientpositive/transform_ppr2.q.out | 24 - .../truncate_column_list_bucket.q.out | 22 - .../results/clientpositive/udf_explode.q.out | 32 - .../results/clientpositive/udtf_explode.q.out | 32 - .../test/results/clientpositive/union22.q.out | 81 --- .../test/results/clientpositive/union24.q.out | 192 ------ .../results/clientpositive/union_ppr.q.out | 24 - .../serde2/dynamic_type/DynamicSerDe.java | 83 +++ .../hive/metastore/utils/MetaStoreUtils.java | 22 +- 200 files changed, 98 insertions(+), 14677 deletions(-) diff --git a/itests/hive-blobstore/src/test/results/clientpositive/insert_into_dynamic_partitions.q.out b/itests/hive-blobstore/src/test/results/clientpositive/insert_into_dynamic_partitions.q.out index 3783c15203..fd82a0c0f8 100644 --- a/itests/hive-blobstore/src/test/results/clientpositive/insert_into_dynamic_partitions.q.out +++ b/itests/hive-blobstore/src/test/results/clientpositive/insert_into_dynamic_partitions.q.out @@ -129,11 +129,9 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns - columns.comments columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -141,7 +139,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns @@ -149,7 +146,6 @@ STAGE PLANS: columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -187,10 +183,8 @@ STAGE PLANS: name default.table1 partition_columns key partition_columns.types string - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 TotalFiles: 1 @@ -220,10 +214,8 @@ STAGE PLANS: name default.table1 partition_columns key partition_columns.types string - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 diff --git a/itests/hive-blobstore/src/test/results/clientpositive/insert_into_table.q.out b/itests/hive-blobstore/src/test/results/clientpositive/insert_into_table.q.out index 92c785c15a..ce071cba37 100644 --- a/itests/hive-blobstore/src/test/results/clientpositive/insert_into_table.q.out +++ b/itests/hive-blobstore/src/test/results/clientpositive/insert_into_table.q.out @@ -90,8 +90,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -100,14 +98,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 2 - numRows 2 - rawDataSize 2 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 TotalFiles: 1 @@ -145,11 +137,9 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns - columns.comments columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -157,7 +147,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns @@ -165,7 +154,6 @@ STAGE PLANS: columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -217,8 +205,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -227,14 +213,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 2 - numRows 2 - rawDataSize 2 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 @@ -263,8 +243,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -273,14 +251,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 2 - numRows 2 - rawDataSize 2 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 TotalFiles: 1 @@ -295,8 +267,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -305,21 +275,13 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 2 - numRows 2 - rawDataSize 2 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -328,14 +290,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 2 - numRows 2 - rawDataSize 2 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 name: default.table1 @@ -351,8 +307,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -361,14 +315,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 2 - numRows 2 - rawDataSize 2 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 @@ -387,8 +335,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -397,14 +343,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 2 - numRows 2 - rawDataSize 2 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 TotalFiles: 1 @@ -419,8 +359,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -429,21 +367,13 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 2 - numRows 2 - rawDataSize 2 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -452,14 +382,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 2 - numRows 2 - rawDataSize 2 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 name: default.table1 diff --git a/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_directory.q.out b/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_directory.q.out index a113a225ba..2eec9f2d78 100644 --- a/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_directory.q.out +++ b/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_directory.q.out @@ -120,31 +120,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns id,key - columns.comments columns.types int:string field.delim , #### A masked pattern was here #### name default.table1 - numFiles 2 - numRows 2 - rawDataSize 8 - serialization.ddl struct table1 { i32 id, string key} serialization.format , serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 10 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id,key @@ -153,14 +143,8 @@ STAGE PLANS: field.delim , #### A masked pattern was here #### name default.table1 - numFiles 2 - numRows 2 - rawDataSize 8 - serialization.ddl struct table1 { i32 id, string key} serialization.format , serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 10 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 name: default.table1 @@ -392,31 +376,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns id,key - columns.comments columns.types int:string field.delim , #### A masked pattern was here #### name default.table1 - numFiles 2 - numRows 2 - rawDataSize 8 - serialization.ddl struct table1 { i32 id, string key} serialization.format , serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 10 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id,key @@ -425,14 +399,8 @@ STAGE PLANS: field.delim , #### A masked pattern was here #### name default.table1 - numFiles 2 - numRows 2 - rawDataSize 8 - serialization.ddl struct table1 { i32 id, string key} serialization.format , serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 10 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 name: default.table1 diff --git a/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_dynamic_partitions.q.out b/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_dynamic_partitions.q.out index 91e95c4213..bdd82d4836 100644 --- a/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_dynamic_partitions.q.out +++ b/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_dynamic_partitions.q.out @@ -147,11 +147,9 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns - columns.comments columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -159,7 +157,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns @@ -167,7 +164,6 @@ STAGE PLANS: columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -205,10 +201,8 @@ STAGE PLANS: name default.table1 partition_columns key partition_columns.types string - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 TotalFiles: 1 @@ -238,10 +232,8 @@ STAGE PLANS: name default.table1 partition_columns key partition_columns.types string - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 diff --git a/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_table.q.out b/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_table.q.out index 96e77ed625..f65bf22eda 100644 --- a/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_table.q.out +++ b/itests/hive-blobstore/src/test/results/clientpositive/insert_overwrite_table.q.out @@ -98,8 +98,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -108,14 +106,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 1 - numRows 1 - rawDataSize 1 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 TotalFiles: 1 @@ -153,11 +145,9 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns - columns.comments columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -165,7 +155,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns @@ -173,7 +162,6 @@ STAGE PLANS: columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -225,8 +213,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -235,14 +221,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 1 - numRows 1 - rawDataSize 1 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 @@ -271,8 +251,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -281,14 +259,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 1 - numRows 1 - rawDataSize 1 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 TotalFiles: 1 @@ -303,8 +275,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -313,21 +283,13 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 1 - numRows 1 - rawDataSize 1 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -336,14 +298,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 1 - numRows 1 - rawDataSize 1 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 name: default.table1 @@ -359,8 +315,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -369,14 +323,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 1 - numRows 1 - rawDataSize 1 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 @@ -395,8 +343,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -405,14 +351,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 1 - numRows 1 - rawDataSize 1 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 TotalFiles: 1 @@ -427,8 +367,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -437,21 +375,13 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 1 - numRows 1 - rawDataSize 1 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns id @@ -460,14 +390,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/table1 name default.table1 - numFiles 1 - numRows 1 - rawDataSize 1 - serialization.ddl struct table1 { i32 id} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 name: default.table1 diff --git a/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out b/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out index 2dd98ef438..8bfc66795e 100644 --- a/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out +++ b/itests/hive-blobstore/src/test/results/clientpositive/write_final_output_blobstore.q.out @@ -80,30 +80,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key - columns.comments columns.types int #### A masked pattern was here #### name default.hdfs_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct hdfs_table { i32 key} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key @@ -111,14 +101,8 @@ STAGE PLANS: columns.types int #### A masked pattern was here #### name default.hdfs_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct hdfs_table { i32 key} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.hdfs_table name: default.hdfs_table @@ -211,8 +195,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key @@ -221,14 +203,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/write_final_output_blobstore name default.blobstore_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct blobstore_table { i32 key} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.blobstore_table TotalFiles: 1 @@ -281,8 +257,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key @@ -291,14 +265,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/write_final_output_blobstore name default.blobstore_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct blobstore_table { i32 key} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.blobstore_table @@ -368,30 +336,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key - columns.comments columns.types int #### A masked pattern was here #### name default.hdfs_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct hdfs_table { i32 key} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key @@ -399,14 +357,8 @@ STAGE PLANS: columns.types int #### A masked pattern was here #### name default.hdfs_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct hdfs_table { i32 key} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.hdfs_table name: default.hdfs_table @@ -499,8 +451,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key @@ -509,14 +459,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/write_final_output_blobstore name default.blobstore_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct blobstore_table { i32 key} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.blobstore_table TotalFiles: 1 @@ -569,8 +513,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key @@ -579,14 +521,8 @@ STAGE PLANS: #### A masked pattern was here #### location ### test.blobstore.path ###/write_final_output_blobstore name default.blobstore_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct blobstore_table { i32 key} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.blobstore_table diff --git a/ql/src/java/org/apache/hadoop/hive/ql/plan/PartitionDesc.java b/ql/src/java/org/apache/hadoop/hive/ql/plan/PartitionDesc.java index b226ab7b24..fda8f461e1 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/plan/PartitionDesc.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/plan/PartitionDesc.java @@ -73,8 +73,6 @@ public void setBaseFileName(String baseFileName) { public PartitionDesc() { } - private final static org.slf4j.Logger LOG = org.slf4j.LoggerFactory.getLogger(PartitionDesc.class); - public PartitionDesc(final TableDesc table, final LinkedHashMap partSpec) { this.tableDesc = table; setPartSpec(partSpec); @@ -92,10 +90,6 @@ public PartitionDesc(final Partition part, final TableDesc tableDesc) throws Hiv } } - public PartitionDesc(final Partition part) throws HiveException { - this(part, getTableDesc(part.getTable())); - } - /** * @param part Partition * @param tblDesc Table Descriptor @@ -223,6 +217,7 @@ public Map getPropertiesExplain() { } public void setProperties(final Properties properties) { + properties.remove("columns.comments"); if (properties instanceof CopyOnFirstWriteProperties) { this.properties = properties; } else { diff --git a/ql/src/java/org/apache/hadoop/hive/ql/plan/TableDesc.java b/ql/src/java/org/apache/hadoop/hive/ql/plan/TableDesc.java index 7993779562..0435477ce4 100644 --- a/ql/src/java/org/apache/hadoop/hive/ql/plan/TableDesc.java +++ b/ql/src/java/org/apache/hadoop/hive/ql/plan/TableDesc.java @@ -45,8 +45,6 @@ */ public class TableDesc implements Serializable, Cloneable { - private static final Logger LOG = LoggerFactory.getLogger(TableDesc.class); - private static final long serialVersionUID = 1L; private Class inputFileFormatClass; private Class outputFileFormatClass; diff --git a/ql/src/test/results/clientpositive/beeline/smb_mapjoin_11.q.out b/ql/src/test/results/clientpositive/beeline/smb_mapjoin_11.q.out index 40895970a9..5d09134eb8 100644 --- a/ql/src/test/results/clientpositive/beeline/smb_mapjoin_11.q.out +++ b/ql/src/test/results/clientpositive/beeline/smb_mapjoin_11.q.out @@ -111,25 +111,17 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 16 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.test_table1_n1 - numFiles 16 - numRows 500 partition_columns ds partition_columns.types string - rawDataSize 5312 - serialization.ddl struct test_table1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -147,10 +139,8 @@ STAGE PLANS: name default.test_table1_n1 partition_columns ds partition_columns.types string - serialization.ddl struct test_table1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table1_n1 name: default.test_table1_n1 @@ -186,10 +176,8 @@ STAGE PLANS: name default.test_table3_n1 partition_columns ds partition_columns.types string - serialization.ddl struct test_table3_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table3_n1 TotalFiles: 16 @@ -248,10 +236,8 @@ STAGE PLANS: name default.test_table3_n1 partition_columns ds partition_columns.types string - serialization.ddl struct test_table3_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table3_n1 @@ -1977,25 +1963,17 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 16 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.test_table1_n1 - numFiles 16 - numRows 500 partition_columns ds partition_columns.types string - rawDataSize 5312 - serialization.ddl struct test_table1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -2013,10 +1991,8 @@ STAGE PLANS: name default.test_table1_n1 partition_columns ds partition_columns.types string - serialization.ddl struct test_table1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table1_n1 name: default.test_table1_n1 @@ -2096,25 +2072,17 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 16 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.test_table3_n1 - numFiles 16 - numRows 1028 partition_columns ds partition_columns.types string - rawDataSize 10968 - serialization.ddl struct test_table3_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11996 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -2131,10 +2099,8 @@ STAGE PLANS: name default.test_table3_n1 partition_columns ds partition_columns.types string - serialization.ddl struct test_table3_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table3_n1 name: default.test_table3_n1 diff --git a/ql/src/test/results/clientpositive/beeline/smb_mapjoin_12.q.out b/ql/src/test/results/clientpositive/beeline/smb_mapjoin_12.q.out index d7bd71aef8..bbe6b8baed 100644 --- a/ql/src/test/results/clientpositive/beeline/smb_mapjoin_12.q.out +++ b/ql/src/test/results/clientpositive/beeline/smb_mapjoin_12.q.out @@ -129,10 +129,8 @@ STAGE PLANS: name default.test_table3_n4 partition_columns ds partition_columns.types string - serialization.ddl struct test_table3_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table3_n4 TotalFiles: 1 @@ -149,25 +147,17 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 16 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.test_table1_n6 - numFiles 16 - numRows 500 partition_columns ds partition_columns.types string - rawDataSize 5312 - serialization.ddl struct test_table1_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -185,10 +175,8 @@ STAGE PLANS: name default.test_table1_n6 partition_columns ds partition_columns.types string - serialization.ddl struct test_table1_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table1_n6 name: default.test_table1_n6 @@ -218,10 +206,8 @@ STAGE PLANS: name default.test_table3_n4 partition_columns ds partition_columns.types string - serialization.ddl struct test_table3_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table3_n4 @@ -343,25 +329,17 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 16 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.test_table3_n4 - numFiles 16 - numRows 3084 partition_columns ds partition_columns.types string - rawDataSize 32904 - serialization.ddl struct test_table3_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 35988 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -379,10 +357,8 @@ STAGE PLANS: name default.test_table3_n4 partition_columns ds partition_columns.types string - serialization.ddl struct test_table3_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table3_n4 name: default.test_table3_n4 @@ -419,10 +395,8 @@ STAGE PLANS: name default.test_table3_n4 partition_columns ds partition_columns.types string - serialization.ddl struct test_table3_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table3_n4 TotalFiles: 16 @@ -482,10 +456,8 @@ STAGE PLANS: name default.test_table3_n4 partition_columns ds partition_columns.types string - serialization.ddl struct test_table3_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table3_n4 diff --git a/ql/src/test/results/clientpositive/beeline/smb_mapjoin_13.q.out b/ql/src/test/results/clientpositive/beeline/smb_mapjoin_13.q.out index 70033d814c..418492077e 100644 --- a/ql/src/test/results/clientpositive/beeline/smb_mapjoin_13.q.out +++ b/ql/src/test/results/clientpositive/beeline/smb_mapjoin_13.q.out @@ -124,31 +124,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.test_table1_n12 - numFiles 16 - numRows 500 - rawDataSize 5312 - serialization.ddl struct test_table1_n12 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key @@ -159,14 +150,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.test_table1_n12 - numFiles 16 - numRows 500 - rawDataSize 5312 - serialization.ddl struct test_table1_n12 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table1_n12 name: default.test_table1_n12 @@ -322,31 +307,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key,value bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.test_table3_n6 - numFiles 16 - numRows 500 - rawDataSize 5312 - serialization.ddl struct test_table3_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key,value @@ -357,14 +333,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.test_table3_n6 - numFiles 16 - numRows 500 - rawDataSize 5312 - serialization.ddl struct test_table3_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table3_n6 name: default.test_table3_n6 diff --git a/ql/src/test/results/clientpositive/binary_output_format.q.out b/ql/src/test/results/clientpositive/binary_output_format.q.out index b414360855..760d86398f 100644 --- a/ql/src/test/results/clientpositive/binary_output_format.q.out +++ b/ql/src/test/results/clientpositive/binary_output_format.q.out @@ -101,8 +101,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns mydata @@ -110,15 +108,9 @@ STAGE PLANS: columns.types string #### A masked pattern was here #### name default.dest1_n109 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n109 { string mydata} serialization.format 1 serialization.last.column.takes.rest true serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n109 TotalFiles: 1 @@ -152,30 +144,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -183,14 +165,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -246,8 +222,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns mydata @@ -255,15 +229,9 @@ STAGE PLANS: columns.types string #### A masked pattern was here #### name default.dest1_n109 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n109 { string mydata} serialization.format 1 serialization.last.column.takes.rest true serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n109 @@ -292,8 +260,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns mydata @@ -301,15 +267,9 @@ STAGE PLANS: columns.types string #### A masked pattern was here #### name default.dest1_n109 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n109 { string mydata} serialization.format 1 serialization.last.column.takes.rest true serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n109 TotalFiles: 1 @@ -324,8 +284,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns mydata @@ -333,22 +291,14 @@ STAGE PLANS: columns.types string #### A masked pattern was here #### name default.dest1_n109 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n109 { string mydata} serialization.format 1 serialization.last.column.takes.rest true serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns mydata @@ -356,15 +306,9 @@ STAGE PLANS: columns.types string #### A masked pattern was here #### name default.dest1_n109 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n109 { string mydata} serialization.format 1 serialization.last.column.takes.rest true serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n109 name: default.dest1_n109 @@ -386,8 +330,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns mydata @@ -395,15 +337,9 @@ STAGE PLANS: columns.types string #### A masked pattern was here #### name default.dest1_n109 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n109 { string mydata} serialization.format 1 serialization.last.column.takes.rest true serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n109 TotalFiles: 1 @@ -418,8 +354,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns mydata @@ -427,22 +361,14 @@ STAGE PLANS: columns.types string #### A masked pattern was here #### name default.dest1_n109 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n109 { string mydata} serialization.format 1 serialization.last.column.takes.rest true serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveBinaryOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"mydata":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns mydata @@ -450,15 +376,9 @@ STAGE PLANS: columns.types string #### A masked pattern was here #### name default.dest1_n109 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n109 { string mydata} serialization.format 1 serialization.last.column.takes.rest true serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n109 name: default.dest1_n109 diff --git a/ql/src/test/results/clientpositive/bucket_map_join_1.q.out b/ql/src/test/results/clientpositive/bucket_map_join_1.q.out index 440345fa53..9d72125cbd 100644 --- a/ql/src/test/results/clientpositive/bucket_map_join_1.q.out +++ b/ql/src/test/results/clientpositive/bucket_map_join_1.q.out @@ -136,18 +136,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.table1_n9 - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct table1_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -163,14 +156,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.table1_n9 - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct table1_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1_n9 name: default.table1_n9 diff --git a/ql/src/test/results/clientpositive/bucket_map_join_2.q.out b/ql/src/test/results/clientpositive/bucket_map_join_2.q.out index 20b27033b8..6e7414c98b 100644 --- a/ql/src/test/results/clientpositive/bucket_map_join_2.q.out +++ b/ql/src/test/results/clientpositive/bucket_map_join_2.q.out @@ -136,18 +136,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.table1 - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct table1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -163,14 +156,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.table1 - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct table1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table1 name: default.table1 diff --git a/ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out b/ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out index 98a45ee2fd..83516bf2f5 100644 --- a/ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out +++ b/ql/src/test/results/clientpositive/bucket_map_join_spark1.q.out @@ -159,20 +159,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n19 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -189,10 +182,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n19 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n19 name: default.srcbucket_mapjoin_part_n19 @@ -258,8 +249,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -267,14 +256,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n9 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n9 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n9 TotalFiles: 1 @@ -326,20 +309,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n16 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n16 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -356,10 +332,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n16 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n16 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n16 name: default.srcbucket_mapjoin_part_2_n16 @@ -375,20 +349,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n19 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -405,10 +372,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n19 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n19 name: default.srcbucket_mapjoin_part_n19 @@ -424,8 +389,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -433,14 +396,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n9 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n9 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n9 @@ -627,20 +584,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n19 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -657,10 +607,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n19 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n19 name: default.srcbucket_mapjoin_part_n19 @@ -726,8 +674,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -735,14 +681,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n9 - numFiles 1 - numRows 1028 - rawDataSize 19022 - serialization.ddl struct bucketmapjoin_tmp_result_n9 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20050 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n9 TotalFiles: 1 @@ -794,20 +734,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n16 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n16 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -824,10 +757,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n16 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n16 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n16 name: default.srcbucket_mapjoin_part_2_n16 @@ -843,20 +774,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n19 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -873,10 +797,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n19 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n19 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n19 name: default.srcbucket_mapjoin_part_n19 @@ -892,8 +814,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -901,14 +821,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n9 - numFiles 1 - numRows 1028 - rawDataSize 19022 - serialization.ddl struct bucketmapjoin_tmp_result_n9 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20050 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n9 diff --git a/ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out b/ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out index 902c1291f3..c95804c623 100644 --- a/ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out +++ b/ql/src/test/results/clientpositive/bucket_map_join_spark2.q.out @@ -143,20 +143,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n10 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -173,10 +166,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n10 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n10 name: default.srcbucket_mapjoin_part_2_n10 @@ -242,8 +233,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -251,14 +240,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n5 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n5 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n5 TotalFiles: 1 @@ -310,20 +293,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n10 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -340,10 +316,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n10 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n10 name: default.srcbucket_mapjoin_part_2_n10 @@ -359,20 +333,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n12 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n12 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -389,10 +356,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n12 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n12 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n12 name: default.srcbucket_mapjoin_part_n12 @@ -408,8 +373,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -417,14 +380,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n5 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n5 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n5 @@ -611,20 +568,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n10 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -641,10 +591,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n10 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n10 name: default.srcbucket_mapjoin_part_2_n10 @@ -710,8 +658,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -719,14 +665,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n5 - numFiles 1 - numRows 564 - rawDataSize 10503 - serialization.ddl struct bucketmapjoin_tmp_result_n5 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11067 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n5 TotalFiles: 1 @@ -778,20 +718,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n10 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -808,10 +741,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n10 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n10 name: default.srcbucket_mapjoin_part_2_n10 @@ -827,20 +758,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n12 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n12 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -857,10 +781,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n12 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n12 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n12 name: default.srcbucket_mapjoin_part_n12 @@ -876,8 +798,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -885,14 +805,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n5 - numFiles 1 - numRows 564 - rawDataSize 10503 - serialization.ddl struct bucketmapjoin_tmp_result_n5 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11067 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n5 diff --git a/ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out b/ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out index 42a6998e95..bbff481863 100644 --- a/ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out +++ b/ql/src/test/results/clientpositive/bucket_map_join_spark3.q.out @@ -143,20 +143,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n4 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -173,10 +166,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n4 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n4 name: default.srcbucket_mapjoin_part_n4 @@ -242,8 +233,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -251,14 +240,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n1 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n1 TotalFiles: 1 @@ -310,20 +293,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n3 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -340,10 +316,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n3 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n3 name: default.srcbucket_mapjoin_part_2_n3 @@ -359,20 +333,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n4 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -389,10 +356,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n4 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n4 name: default.srcbucket_mapjoin_part_n4 @@ -408,8 +373,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -417,14 +380,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n1 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n1 @@ -611,20 +568,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n4 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -641,10 +591,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n4 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n4 name: default.srcbucket_mapjoin_part_n4 @@ -710,8 +658,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -719,14 +665,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n1 - numFiles 1 - numRows 564 - rawDataSize 10503 - serialization.ddl struct bucketmapjoin_tmp_result_n1 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11067 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n1 TotalFiles: 1 @@ -778,20 +718,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n3 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -808,10 +741,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n3 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n3 name: default.srcbucket_mapjoin_part_2_n3 @@ -827,20 +758,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n4 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -857,10 +781,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n4 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n4 name: default.srcbucket_mapjoin_part_n4 @@ -876,8 +798,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -885,14 +805,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n1 - numFiles 1 - numRows 564 - rawDataSize 10503 - serialization.ddl struct bucketmapjoin_tmp_result_n1 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11067 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n1 diff --git a/ql/src/test/results/clientpositive/bucket_map_join_spark4.q.out b/ql/src/test/results/clientpositive/bucket_map_join_spark4.q.out index 5e6a28b4e8..1528a78dbd 100644 --- a/ql/src/test/results/clientpositive/bucket_map_join_spark4.q.out +++ b/ql/src/test/results/clientpositive/bucket_map_join_spark4.q.out @@ -212,31 +212,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.tbl1_n0 - numFiles 2 - numRows 10 - rawDataSize 70 - serialization.ddl struct tbl1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 80 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -247,14 +238,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.tbl1_n0 - numFiles 2 - numRows 10 - rawDataSize 70 - serialization.ddl struct tbl1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 80 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tbl1_n0 name: default.tbl1_n0 @@ -264,31 +249,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.tbl2_n0 - numFiles 2 - numRows 10 - rawDataSize 70 - serialization.ddl struct tbl2_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 80 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -299,14 +275,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.tbl2_n0 - numFiles 2 - numRows 10 - rawDataSize 70 - serialization.ddl struct tbl2_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 80 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tbl2_n0 name: default.tbl2_n0 @@ -316,31 +286,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.tbl3 - numFiles 2 - numRows 10 - rawDataSize 70 - serialization.ddl struct tbl3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 80 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -351,14 +312,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.tbl3 - numFiles 2 - numRows 10 - rawDataSize 70 - serialization.ddl struct tbl3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 80 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tbl3 name: default.tbl3 @@ -597,31 +552,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.tbl1_n0 - numFiles 2 - numRows 10 - rawDataSize 70 - serialization.ddl struct tbl1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 80 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -632,14 +578,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.tbl1_n0 - numFiles 2 - numRows 10 - rawDataSize 70 - serialization.ddl struct tbl1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 80 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tbl1_n0 name: default.tbl1_n0 @@ -649,31 +589,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.tbl2_n0 - numFiles 2 - numRows 10 - rawDataSize 70 - serialization.ddl struct tbl2_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 80 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -684,14 +615,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.tbl2_n0 - numFiles 2 - numRows 10 - rawDataSize 70 - serialization.ddl struct tbl2_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 80 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tbl2_n0 name: default.tbl2_n0 @@ -701,31 +626,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.tbl3 - numFiles 2 - numRows 10 - rawDataSize 70 - serialization.ddl struct tbl3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 80 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -736,14 +652,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.tbl3 - numFiles 2 - numRows 10 - rawDataSize 70 - serialization.ddl struct tbl3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 80 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tbl3 name: default.tbl3 diff --git a/ql/src/test/results/clientpositive/bucketcontext_1.q.out b/ql/src/test/results/clientpositive/bucketcontext_1.q.out index 35b6ae89ca..515ec2b97b 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_1.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_1.q.out @@ -140,20 +140,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n14 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n14 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -171,10 +164,8 @@ STAGE PLANS: name default.bucket_small_n14 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n14 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n14 name: default.bucket_small_n14 @@ -256,20 +247,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n14 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n14 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -287,10 +271,8 @@ STAGE PLANS: name default.bucket_big_n14 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n14 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n14 name: default.bucket_big_n14 @@ -306,20 +288,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n14 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n14 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -337,10 +312,8 @@ STAGE PLANS: name default.bucket_big_n14 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n14 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n14 name: default.bucket_big_n14 @@ -474,20 +447,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n14 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n14 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -505,10 +471,8 @@ STAGE PLANS: name default.bucket_big_n14 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n14 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n14 name: default.bucket_big_n14 @@ -524,20 +488,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n14 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n14 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -555,10 +512,8 @@ STAGE PLANS: name default.bucket_big_n14 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n14 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n14 name: default.bucket_big_n14 diff --git a/ql/src/test/results/clientpositive/bucketcontext_2.q.out b/ql/src/test/results/clientpositive/bucketcontext_2.q.out index 442e93b778..6c5c25148d 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_2.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_2.q.out @@ -124,20 +124,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n8 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -155,10 +148,8 @@ STAGE PLANS: name default.bucket_small_n8 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n8 name: default.bucket_small_n8 @@ -240,20 +231,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n8 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -271,10 +255,8 @@ STAGE PLANS: name default.bucket_big_n8 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n8 name: default.bucket_big_n8 @@ -290,20 +272,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n8 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -321,10 +296,8 @@ STAGE PLANS: name default.bucket_big_n8 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n8 name: default.bucket_big_n8 @@ -458,20 +431,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n8 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -489,10 +455,8 @@ STAGE PLANS: name default.bucket_big_n8 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n8 name: default.bucket_big_n8 @@ -508,20 +472,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n8 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -539,10 +496,8 @@ STAGE PLANS: name default.bucket_big_n8 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n8 name: default.bucket_big_n8 diff --git a/ql/src/test/results/clientpositive/bucketcontext_3.q.out b/ql/src/test/results/clientpositive/bucketcontext_3.q.out index 8fa490d68d..6358b46d39 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_3.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_3.q.out @@ -124,20 +124,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n4 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -155,10 +148,8 @@ STAGE PLANS: name default.bucket_small_n4 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n4 name: default.bucket_small_n4 @@ -172,20 +163,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n4 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -203,10 +187,8 @@ STAGE PLANS: name default.bucket_small_n4 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n4 name: default.bucket_small_n4 @@ -288,20 +270,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n4 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -319,10 +294,8 @@ STAGE PLANS: name default.bucket_big_n4 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n4 name: default.bucket_big_n4 @@ -455,20 +428,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n4 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -486,10 +452,8 @@ STAGE PLANS: name default.bucket_big_n4 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n4 name: default.bucket_big_n4 diff --git a/ql/src/test/results/clientpositive/bucketcontext_4.q.out b/ql/src/test/results/clientpositive/bucketcontext_4.q.out index c0f1017536..74121e01d7 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_4.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_4.q.out @@ -140,20 +140,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -171,10 +164,8 @@ STAGE PLANS: name default.bucket_small partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small name: default.bucket_small @@ -188,20 +179,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -219,10 +203,8 @@ STAGE PLANS: name default.bucket_small partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small name: default.bucket_small @@ -304,20 +286,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -335,10 +310,8 @@ STAGE PLANS: name default.bucket_big partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big name: default.bucket_big @@ -471,20 +444,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -502,10 +468,8 @@ STAGE PLANS: name default.bucket_big partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big name: default.bucket_big diff --git a/ql/src/test/results/clientpositive/bucketcontext_5.q.out b/ql/src/test/results/clientpositive/bucketcontext_5.q.out index ac4496cf2f..066adfd447 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_5.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_5.q.out @@ -166,18 +166,11 @@ STAGE PLANS: bucketing_version 1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n13 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_big_n13 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -193,14 +186,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big_n13 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_big_n13 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n13 name: default.bucket_big_n13 @@ -321,18 +308,11 @@ STAGE PLANS: bucketing_version 1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n13 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_big_n13 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -348,14 +328,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big_n13 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_big_n13 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n13 name: default.bucket_big_n13 diff --git a/ql/src/test/results/clientpositive/bucketcontext_6.q.out b/ql/src/test/results/clientpositive/bucketcontext_6.q.out index 39ed5cce50..5147d67ec0 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_6.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_6.q.out @@ -188,20 +188,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n7 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -219,10 +212,8 @@ STAGE PLANS: name default.bucket_big_n7 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n7 name: default.bucket_big_n7 @@ -238,20 +229,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n7 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -269,10 +253,8 @@ STAGE PLANS: name default.bucket_big_n7 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n7 name: default.bucket_big_n7 @@ -402,20 +384,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n7 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -433,10 +408,8 @@ STAGE PLANS: name default.bucket_big_n7 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n7 name: default.bucket_big_n7 @@ -452,20 +425,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n7 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -483,10 +449,8 @@ STAGE PLANS: name default.bucket_big_n7 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n7 name: default.bucket_big_n7 diff --git a/ql/src/test/results/clientpositive/bucketcontext_7.q.out b/ql/src/test/results/clientpositive/bucketcontext_7.q.out index eb64514c9e..f488e4128e 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_7.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_7.q.out @@ -159,20 +159,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n2 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -190,10 +183,8 @@ STAGE PLANS: name default.bucket_small_n2 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n2 name: default.bucket_small_n2 @@ -207,20 +198,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n2 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -238,10 +222,8 @@ STAGE PLANS: name default.bucket_small_n2 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n2 name: default.bucket_small_n2 @@ -323,20 +305,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n2 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -354,10 +329,8 @@ STAGE PLANS: name default.bucket_big_n2 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n2 name: default.bucket_big_n2 @@ -373,20 +346,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n2 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -404,10 +370,8 @@ STAGE PLANS: name default.bucket_big_n2 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n2 name: default.bucket_big_n2 @@ -545,20 +509,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n2 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -576,10 +533,8 @@ STAGE PLANS: name default.bucket_big_n2 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n2 name: default.bucket_big_n2 @@ -595,20 +550,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n2 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -626,10 +574,8 @@ STAGE PLANS: name default.bucket_big_n2 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n2 name: default.bucket_big_n2 diff --git a/ql/src/test/results/clientpositive/bucketcontext_8.q.out b/ql/src/test/results/clientpositive/bucketcontext_8.q.out index 245b9618ea..17759470a6 100644 --- a/ql/src/test/results/clientpositive/bucketcontext_8.q.out +++ b/ql/src/test/results/clientpositive/bucketcontext_8.q.out @@ -159,20 +159,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n10 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n10 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -190,10 +183,8 @@ STAGE PLANS: name default.bucket_small_n10 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n10 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n10 name: default.bucket_small_n10 @@ -207,20 +198,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n10 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n10 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -238,10 +222,8 @@ STAGE PLANS: name default.bucket_small_n10 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n10 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n10 name: default.bucket_small_n10 @@ -323,20 +305,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n10 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n10 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -354,10 +329,8 @@ STAGE PLANS: name default.bucket_big_n10 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n10 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n10 name: default.bucket_big_n10 @@ -373,20 +346,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n10 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n10 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -404,10 +370,8 @@ STAGE PLANS: name default.bucket_big_n10 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n10 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n10 name: default.bucket_big_n10 @@ -545,20 +509,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n10 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n10 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -576,10 +533,8 @@ STAGE PLANS: name default.bucket_big_n10 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n10 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n10 name: default.bucket_big_n10 @@ -595,20 +550,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n10 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n10 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -626,10 +574,8 @@ STAGE PLANS: name default.bucket_big_n10 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n10 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n10 name: default.bucket_big_n10 diff --git a/ql/src/test/results/clientpositive/bucketmapjoin10.q.out b/ql/src/test/results/clientpositive/bucketmapjoin10.q.out index c4d607054f..c57f403d7b 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin10.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin10.q.out @@ -173,20 +173,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n13 - numFiles 3 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n13 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -203,10 +196,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n13 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n13 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n13 name: default.srcbucket_mapjoin_part_2_n13 @@ -220,20 +211,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n13 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n13 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -250,10 +234,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n13 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n13 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n13 name: default.srcbucket_mapjoin_part_2_n13 @@ -326,20 +308,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n6 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -356,10 +331,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n6 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n6 name: default.srcbucket_mapjoin_part_1_n6 @@ -375,20 +348,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n6 - numFiles 3 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -405,10 +371,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n6 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n6 name: default.srcbucket_mapjoin_part_1_n6 diff --git a/ql/src/test/results/clientpositive/bucketmapjoin11.q.out b/ql/src/test/results/clientpositive/bucketmapjoin11.q.out index 16114c9549..092b605d5c 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin11.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin11.q.out @@ -181,20 +181,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n6 - numFiles 4 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -211,10 +204,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n6 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n6 name: default.srcbucket_mapjoin_part_2_n6 @@ -228,20 +219,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n6 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -258,10 +242,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n6 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n6 name: default.srcbucket_mapjoin_part_2_n6 @@ -342,20 +324,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n2 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -372,10 +347,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n2 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n2 name: default.srcbucket_mapjoin_part_1_n2 @@ -391,20 +364,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n2 - numFiles 4 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -421,10 +387,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n2 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n2 name: default.srcbucket_mapjoin_part_1_n2 @@ -539,20 +503,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n6 - numFiles 4 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -569,10 +526,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n6 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n6 name: default.srcbucket_mapjoin_part_2_n6 @@ -586,20 +541,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n6 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -616,10 +564,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n6 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n6 name: default.srcbucket_mapjoin_part_2_n6 @@ -700,20 +646,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n2 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -730,10 +669,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n2 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n2 name: default.srcbucket_mapjoin_part_1_n2 @@ -749,20 +686,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n2 - numFiles 4 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -779,10 +709,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n2 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n2 name: default.srcbucket_mapjoin_part_1_n2 diff --git a/ql/src/test/results/clientpositive/bucketmapjoin12.q.out b/ql/src/test/results/clientpositive/bucketmapjoin12.q.out index 5c453948cb..8a584b8794 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin12.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin12.q.out @@ -138,26 +138,18 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n0 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -167,10 +159,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n0 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n0 name: default.srcbucket_mapjoin_part_2_n0 @@ -251,20 +241,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -281,10 +264,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1 name: default.srcbucket_mapjoin_part_1 @@ -386,23 +367,15 @@ STAGE PLANS: partition values: part 1 properties: - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_3 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -419,10 +392,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_3 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_3 name: default.srcbucket_mapjoin_part_3 @@ -495,20 +466,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -525,10 +489,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1 name: default.srcbucket_mapjoin_part_1 diff --git a/ql/src/test/results/clientpositive/bucketmapjoin13.q.out b/ql/src/test/results/clientpositive/bucketmapjoin13.q.out index 693377f863..77e986ecdc 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin13.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin13.q.out @@ -103,25 +103,17 @@ STAGE PLANS: partition values: part 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n18 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -138,10 +130,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n18 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n18 name: default.srcbucket_mapjoin_part_2_n18 @@ -210,25 +200,17 @@ STAGE PLANS: partition values: part 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name value column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n8 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -245,10 +227,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n8 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n8 name: default.srcbucket_mapjoin_part_1_n8 @@ -260,25 +240,17 @@ STAGE PLANS: partition values: part 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n8 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -295,10 +267,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n8 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n8 name: default.srcbucket_mapjoin_part_1_n8 @@ -403,25 +373,17 @@ STAGE PLANS: partition values: part 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n18 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -438,10 +400,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n18 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n18 name: default.srcbucket_mapjoin_part_2_n18 @@ -518,25 +478,17 @@ STAGE PLANS: partition values: part 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n8 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -553,10 +505,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n8 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n8 name: default.srcbucket_mapjoin_part_1_n8 @@ -666,25 +616,17 @@ STAGE PLANS: partition values: part 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n18 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -701,10 +643,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n18 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n18 name: default.srcbucket_mapjoin_part_2_n18 @@ -781,25 +721,17 @@ STAGE PLANS: partition values: part 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n8 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -816,10 +748,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n8 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n8 name: default.srcbucket_mapjoin_part_1_n8 @@ -929,25 +859,17 @@ STAGE PLANS: partition values: part 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n18 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -964,10 +886,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n18 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n18 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n18 name: default.srcbucket_mapjoin_part_2_n18 @@ -1044,25 +964,17 @@ STAGE PLANS: partition values: part 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n8 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1079,10 +991,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n8 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n8 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n8 name: default.srcbucket_mapjoin_part_1_n8 diff --git a/ql/src/test/results/clientpositive/bucketmapjoin5.q.out b/ql/src/test/results/clientpositive/bucketmapjoin5.q.out index b0e2931570..b43c6758ed 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin5.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin5.q.out @@ -264,8 +264,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -273,14 +271,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result TotalFiles: 1 @@ -322,20 +314,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n0 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -352,10 +337,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n0 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n0 name: default.srcbucket_mapjoin_part_n0 @@ -371,20 +354,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n0 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -401,10 +377,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n0 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n0 name: default.srcbucket_mapjoin_part_n0 @@ -461,8 +435,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -470,14 +442,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result @@ -506,8 +472,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -515,14 +479,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result TotalFiles: 1 @@ -537,8 +495,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -546,21 +502,13 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -568,14 +516,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result name: default.bucketmapjoin_tmp_result @@ -597,8 +539,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -606,14 +546,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result TotalFiles: 1 @@ -628,8 +562,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -637,21 +569,13 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -659,14 +583,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result name: default.bucketmapjoin_tmp_result @@ -886,8 +804,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -895,14 +811,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 1 - numRows 928 - rawDataSize 17038 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17966 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result TotalFiles: 1 @@ -944,20 +854,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -974,10 +877,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2 name: default.srcbucket_mapjoin_part_2 @@ -993,20 +894,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1023,10 +917,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2 name: default.srcbucket_mapjoin_part_2 @@ -1083,8 +975,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -1092,14 +982,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 1 - numRows 928 - rawDataSize 17038 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17966 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result @@ -1128,8 +1012,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -1137,14 +1019,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 1 - numRows 928 - rawDataSize 17038 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17966 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result TotalFiles: 1 @@ -1159,8 +1035,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -1168,21 +1042,13 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 1 - numRows 928 - rawDataSize 17038 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17966 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -1190,14 +1056,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 1 - numRows 928 - rawDataSize 17038 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17966 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result name: default.bucketmapjoin_tmp_result @@ -1219,8 +1079,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -1228,14 +1086,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 1 - numRows 928 - rawDataSize 17038 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17966 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result TotalFiles: 1 @@ -1250,8 +1102,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -1259,21 +1109,13 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 1 - numRows 928 - rawDataSize 17038 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17966 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -1281,14 +1123,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result - numFiles 1 - numRows 928 - rawDataSize 17038 - serialization.ddl struct bucketmapjoin_tmp_result { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17966 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result name: default.bucketmapjoin_tmp_result diff --git a/ql/src/test/results/clientpositive/bucketmapjoin8.q.out b/ql/src/test/results/clientpositive/bucketmapjoin8.q.out index e1b658e1b4..e70f877801 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin8.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin8.q.out @@ -103,20 +103,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n4 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -133,10 +126,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n4 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n4 name: default.srcbucket_mapjoin_part_2_n4 @@ -217,20 +208,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n1 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -247,10 +231,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n1 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n1 name: default.srcbucket_mapjoin_part_1_n1 @@ -364,20 +346,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n4 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -394,10 +369,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n4 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n4 name: default.srcbucket_mapjoin_part_2_n4 @@ -478,20 +451,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n1 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -508,10 +474,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n1 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n1 name: default.srcbucket_mapjoin_part_1_n1 diff --git a/ql/src/test/results/clientpositive/bucketmapjoin9.q.out b/ql/src/test/results/clientpositive/bucketmapjoin9.q.out index 709c780fd1..49789b5480 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin9.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin9.q.out @@ -111,20 +111,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n12 - numFiles 3 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n12 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -141,10 +134,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n12 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n12 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n12 name: default.srcbucket_mapjoin_part_2_n12 @@ -217,20 +208,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n5 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -247,10 +231,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n5 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n5 name: default.srcbucket_mapjoin_part_1_n5 @@ -397,20 +379,13 @@ STAGE PLANS: bucket_field_name value column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n12 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n12 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -427,10 +402,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n12 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n12 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n12 name: default.srcbucket_mapjoin_part_2_n12 @@ -503,20 +476,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n5 - numFiles 2 - numRows 0 partition_columns part partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -533,10 +499,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n5 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n5 name: default.srcbucket_mapjoin_part_1_n5 diff --git a/ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out b/ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out index c90fa595dc..bd2e6a8cbf 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin_negative.q.out @@ -112,20 +112,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n10 - numFiles 3 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -142,10 +135,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n10 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n10 name: default.srcbucket_mapjoin_part_n10 @@ -203,8 +194,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -212,14 +201,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n4 TotalFiles: 1 @@ -260,18 +243,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n10 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -286,14 +262,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n10 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_n10 name: default.srcbucket_mapjoin_n10 @@ -349,8 +319,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -358,14 +326,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n4 @@ -394,8 +356,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -403,14 +363,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n4 TotalFiles: 1 @@ -425,8 +379,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -434,21 +386,13 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -456,14 +400,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n4 name: default.bucketmapjoin_tmp_result_n4 @@ -485,8 +423,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -494,14 +430,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n4 TotalFiles: 1 @@ -516,8 +446,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -525,21 +453,13 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -547,14 +467,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n4 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n4 name: default.bucketmapjoin_tmp_result_n4 diff --git a/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out b/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out index 1adac2520c..cd3576a652 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin_negative2.q.out @@ -123,20 +123,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n7 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n7 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -153,10 +146,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n7 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n7 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n7 name: default.srcbucket_mapjoin_part_2_n7 @@ -170,20 +161,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n7 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n7 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -200,10 +184,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n7 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n7 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n7 name: default.srcbucket_mapjoin_part_2_n7 @@ -269,8 +251,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -278,14 +258,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n3 TotalFiles: 1 @@ -326,18 +300,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n5 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -352,14 +319,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n5 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_n5 name: default.srcbucket_mapjoin_n5 @@ -415,8 +376,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -424,14 +383,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n3 @@ -460,8 +413,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -469,14 +420,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n3 TotalFiles: 1 @@ -491,8 +436,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -500,21 +443,13 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -522,14 +457,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n3 name: default.bucketmapjoin_tmp_result_n3 @@ -551,8 +480,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -560,14 +487,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n3 TotalFiles: 1 @@ -582,8 +503,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -591,21 +510,13 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -613,14 +524,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n3 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n3 name: default.bucketmapjoin_tmp_result_n3 diff --git a/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out b/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out index 2958ae11ed..3e489ed049 100644 --- a/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out +++ b/ql/src/test/results/clientpositive/bucketmapjoin_negative3.q.out @@ -254,18 +254,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.test1_n11 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test1_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -281,14 +274,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.test1_n11 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test1_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n11 name: default.test1_n11 @@ -413,18 +400,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.test2_n7 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test2_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -440,14 +420,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.test2_n7 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test2_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n7 name: default.test2_n7 @@ -564,18 +538,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.test1_n11 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test1_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -591,14 +558,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.test1_n11 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test1_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n11 name: default.test1_n11 @@ -717,18 +678,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.test1_n11 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test1_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -744,14 +698,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.test1_n11 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test1_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n11 name: default.test1_n11 @@ -870,18 +818,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.test1_n11 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test1_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -897,14 +838,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.test1_n11 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test1_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n11 name: default.test1_n11 @@ -1023,18 +958,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.test1_n11 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test1_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1050,14 +978,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.test1_n11 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test1_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n11 name: default.test1_n11 @@ -1176,18 +1098,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.test2_n7 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test2_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1203,14 +1118,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.test2_n7 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test2_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n7 name: default.test2_n7 @@ -1329,18 +1238,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.test2_n7 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test2_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1356,14 +1258,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.test2_n7 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test2_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n7 name: default.test2_n7 @@ -1482,18 +1378,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.test3 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1509,14 +1398,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.test3 - numFiles 3 - numRows 0 - rawDataSize 0 - serialization.ddl struct test3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4200 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test3 name: default.test3 diff --git a/ql/src/test/results/clientpositive/druid/druidkafkamini_basic.q.out b/ql/src/test/results/clientpositive/druid/druidkafkamini_basic.q.out index 42e9e5279d..8d6ebe989f 100644 --- a/ql/src/test/results/clientpositive/druid/druidkafkamini_basic.q.out +++ b/ql/src/test/results/clientpositive/druid/druidkafkamini_basic.q.out @@ -272,31 +272,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"fullname":"true","shortname":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns shortname,fullname - columns.comments columns.types string:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.languages - numFiles 1 - numRows 2 - rawDataSize 20 - serialization.ddl struct languages { string shortname, string fullname} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"fullname":"true","shortname":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns shortname,fullname @@ -305,14 +295,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.languages - numFiles 1 - numRows 2 - rawDataSize 20 - serialization.ddl struct languages { string shortname, string fullname} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.languages name: default.languages @@ -351,13 +335,11 @@ STAGE PLANS: input format: org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat output format: org.apache.hadoop.hive.druid.io.DruidOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"__time":"true","added":"true","deleted":"true","language":"true","page":"true","user":"true"}} EXTERNAL TRUE bucket_count -1 bucketing_version 2 column.name.delimiter , columns __time,page,user,language,added,deleted - columns.comments columns.types timestamp:string:string:string:int:int druid.datasource default.druid_kafka_test druid.fieldNames language,user @@ -378,23 +360,15 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.druid_kafka_test - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct druid_kafka_test { timestamp __time, string page, string user, string language, i32 added, i32 deleted} serialization.format 1 serialization.lib org.apache.hadoop.hive.druid.serde.DruidSerDe storage_handler org.apache.hadoop.hive.druid.DruidStorageHandler - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.druid.serde.DruidSerDe input format: org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat output format: org.apache.hadoop.hive.druid.io.DruidOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"__time":"true","added":"true","deleted":"true","language":"true","page":"true","user":"true"}} EXTERNAL TRUE - bucket_count -1 bucketing_version 2 column.name.delimiter , columns __time,page,user,language,added,deleted @@ -419,15 +393,9 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.druid_kafka_test - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct druid_kafka_test { timestamp __time, string page, string user, string language, i32 added, i32 deleted} serialization.format 1 serialization.lib org.apache.hadoop.hive.druid.serde.DruidSerDe storage_handler org.apache.hadoop.hive.druid.DruidStorageHandler - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.druid.serde.DruidSerDe name: default.druid_kafka_test name: default.druid_kafka_test diff --git a/ql/src/test/results/clientpositive/druid/druidmini_expressions.q.out b/ql/src/test/results/clientpositive/druid/druidmini_expressions.q.out index a8af291989..67ba37fe98 100644 --- a/ql/src/test/results/clientpositive/druid/druidmini_expressions.q.out +++ b/ql/src/test/results/clientpositive/druid/druidmini_expressions.q.out @@ -240,13 +240,11 @@ STAGE PLANS: input format: org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat output format: org.apache.hadoop.hive.druid.io.DruidOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} EXTERNAL TRUE bucket_count -1 bucketing_version 2 column.name.delimiter , columns __time,cstring1,cstring2,cdouble,cfloat,ctinyint,csmallint,cint,cbigint,cboolean1,cboolean2,cintstring,cfloatstring,cdoublestring - columns.comments columns.types timestamp with local time zone:string:string:double:float:tinyint:smallint:int:bigint:boolean:boolean:string:string:string druid.datasource default.druid_table_alltypesorc druid.fieldNames vc @@ -258,23 +256,15 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.druid_table_alltypesorc - numFiles 0 - numRows 9173 - rawDataSize 0 - serialization.ddl struct druid_table_alltypesorc { timestamp with local time zone __time, string cstring1, string cstring2, double cdouble, float cfloat, byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, bool cboolean1, bool cboolean2, string cintstring, string cfloatstring, string cdoublestring} serialization.format 1 serialization.lib org.apache.hadoop.hive.druid.serde.DruidSerDe storage_handler org.apache.hadoop.hive.druid.DruidStorageHandler - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.druid.serde.DruidSerDe input format: org.apache.hadoop.hive.druid.io.DruidQueryBasedInputFormat output format: org.apache.hadoop.hive.druid.io.DruidOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} EXTERNAL TRUE - bucket_count -1 bucketing_version 2 column.name.delimiter , columns __time,cstring1,cstring2,cdouble,cfloat,ctinyint,csmallint,cint,cbigint,cboolean1,cboolean2,cintstring,cfloatstring,cdoublestring @@ -290,15 +280,9 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.druid_table_alltypesorc - numFiles 0 - numRows 9173 - rawDataSize 0 - serialization.ddl struct druid_table_alltypesorc { timestamp with local time zone __time, string cstring1, string cstring2, double cdouble, float cfloat, byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, bool cboolean1, bool cboolean2, string cintstring, string cfloatstring, string cdoublestring} serialization.format 1 serialization.lib org.apache.hadoop.hive.druid.serde.DruidSerDe storage_handler org.apache.hadoop.hive.druid.DruidStorageHandler - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.druid.serde.DruidSerDe name: default.druid_table_alltypesorc name: default.druid_table_alltypesorc diff --git a/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out b/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out index d6115d84be..894750acbe 100644 --- a/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out +++ b/ql/src/test/results/clientpositive/encrypted/encryption_join_unencrypted_tbl.q.out @@ -615,31 +615,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.encrypted_table - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct encrypted_table { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -648,14 +638,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.encrypted_table - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct encrypted_table { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.encrypted_table name: default.encrypted_table @@ -665,31 +649,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -698,14 +672,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/encrypted/encryption_join_with_different_encryption_keys.q.out b/ql/src/test/results/clientpositive/encrypted/encryption_join_with_different_encryption_keys.q.out index f6c7d50f46..9cd3b0d465 100644 --- a/ql/src/test/results/clientpositive/encrypted/encryption_join_with_different_encryption_keys.q.out +++ b/ql/src/test/results/clientpositive/encrypted/encryption_join_with_different_encryption_keys.q.out @@ -133,31 +133,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.table_key_1 - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct table_key_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -166,14 +156,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.table_key_1 - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct table_key_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table_key_1 name: default.table_key_1 @@ -183,31 +167,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.table_key_2 - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct table_key_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -216,14 +190,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.table_key_2 - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct table_key_2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.table_key_2 name: default.table_key_2 diff --git a/ql/src/test/results/clientpositive/erasurecoding/erasure_explain.q.out b/ql/src/test/results/clientpositive/erasurecoding/erasure_explain.q.out index adad89ef02..c948f8d71f 100644 --- a/ql/src/test/results/clientpositive/erasurecoding/erasure_explain.q.out +++ b/ql/src/test/results/clientpositive/erasurecoding/erasure_explain.q.out @@ -131,32 +131,22 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.srcpart - numFiles 1 numFilesErasureCoded 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -167,10 +157,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -181,32 +169,22 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.srcpart - numFiles 1 numFilesErasureCoded 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -217,10 +195,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -231,32 +207,22 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.srcpart - numFiles 1 numFilesErasureCoded 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -267,10 +233,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -281,32 +245,22 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.srcpart - numFiles 1 numFilesErasureCoded 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -317,10 +271,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/join_map_ppr.q.out b/ql/src/test/results/clientpositive/join_map_ppr.q.out index e3d8212352..bf57f67346 100644 --- a/ql/src/test/results/clientpositive/join_map_ppr.q.out +++ b/ql/src/test/results/clientpositive/join_map_ppr.q.out @@ -124,8 +124,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -133,14 +131,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n4 TotalFiles: 1 @@ -179,30 +171,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -212,10 +194,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -271,8 +251,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -280,14 +258,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n4 @@ -316,8 +288,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -325,14 +295,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n4 TotalFiles: 1 @@ -347,8 +311,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -356,21 +318,13 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -378,14 +332,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n4 name: default.dest_j1_n4 @@ -407,8 +355,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -416,14 +362,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n4 TotalFiles: 1 @@ -438,8 +378,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -447,21 +385,13 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -469,14 +399,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n4 name: default.dest_j1_n4 @@ -783,8 +707,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -792,14 +714,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 1 - numRows 107 - rawDataSize 2018 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2125 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n4 TotalFiles: 1 @@ -838,30 +754,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -871,10 +777,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -930,8 +834,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -939,14 +841,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 1 - numRows 107 - rawDataSize 2018 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2125 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n4 @@ -975,8 +871,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -984,14 +878,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 1 - numRows 107 - rawDataSize 2018 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2125 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n4 TotalFiles: 1 @@ -1006,8 +894,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -1015,21 +901,13 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 1 - numRows 107 - rawDataSize 2018 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2125 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -1037,14 +915,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 1 - numRows 107 - rawDataSize 2018 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2125 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n4 name: default.dest_j1_n4 @@ -1066,8 +938,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -1075,14 +945,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 1 - numRows 107 - rawDataSize 2018 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2125 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n4 TotalFiles: 1 @@ -1097,8 +961,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -1106,21 +968,13 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 1 - numRows 107 - rawDataSize 2018 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2125 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -1128,14 +982,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n4 - numFiles 1 - numRows 107 - rawDataSize 2018 - serialization.ddl struct dest_j1_n4 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2125 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n4 name: default.dest_j1_n4 diff --git a/ql/src/test/results/clientpositive/kafka/kafka_storage_handler.q.out b/ql/src/test/results/clientpositive/kafka/kafka_storage_handler.q.out index 71af39caaa..75fb8230c3 100644 --- a/ql/src/test/results/clientpositive/kafka/kafka_storage_handler.q.out +++ b/ql/src/test/results/clientpositive/kafka/kafka_storage_handler.q.out @@ -1195,7 +1195,6 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns isrobot,channel,timestamp,flags,isunpatrolled,page,diffurl,added,comment,commentlength,isnew,isminor,delta,isanonymous,user,deltabucket,deleted,namespace,__key,__partition,__offset,__timestamp - columns.comments 'from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer' columns.types boolean:string:string:string:boolean:string:string:bigint:string:bigint:boolean:boolean:bigint:boolean:string:double:bigint:string:binary:int:bigint:bigint #### A masked pattern was here #### hive.kafka.max.retries 6 @@ -1209,15 +1208,9 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.wiki_kafka_avro_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct wiki_kafka_avro_table { bool isrobot, string channel, string timestamp, string flags, bool isunpatrolled, string page, string diffurl, i64 added, string comment, i64 commentlength, bool isnew, bool isminor, i64 delta, bool isanonymous, string user, double deltabucket, i64 deleted, string namespace, binary __key, i32 __partition, i64 __offset, i64 __timestamp} serialization.format 1 serialization.lib org.apache.hadoop.hive.kafka.KafkaSerDe storage_handler org.apache.hadoop.hive.kafka.KafkaStorageHandler - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.kafka.KafkaSerDe input format: org.apache.hadoop.hive.kafka.KafkaInputFormat @@ -1294,7 +1287,6 @@ STAGE PLANS: "type" : "string" } ] } - bucket_count -1 bucketing_version 2 column.name.delimiter , columns isrobot,channel,timestamp,flags,isunpatrolled,page,diffurl,added,comment,commentlength,isnew,isminor,delta,isanonymous,user,deltabucket,deleted,namespace,__key,__partition,__offset,__timestamp @@ -1312,15 +1304,9 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.wiki_kafka_avro_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct wiki_kafka_avro_table { bool isrobot, string channel, string timestamp, string flags, bool isunpatrolled, string page, string diffurl, i64 added, string comment, i64 commentlength, bool isnew, bool isminor, i64 delta, bool isanonymous, string user, double deltabucket, i64 deleted, string namespace, binary __key, i32 __partition, i64 __offset, i64 __timestamp} serialization.format 1 serialization.lib org.apache.hadoop.hive.kafka.KafkaSerDe storage_handler org.apache.hadoop.hive.kafka.KafkaStorageHandler - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.kafka.KafkaSerDe name: default.wiki_kafka_avro_table name: default.wiki_kafka_avro_table @@ -1522,7 +1508,6 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns isrobot,channel,timestamp,flags,isunpatrolled,page,diffurl,added,comment,commentlength,isnew,isminor,delta,isanonymous,user,deltabucket,deleted,namespace,__key,__partition,__offset,__timestamp - columns.comments 'from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer','from deserializer' columns.types boolean:string:string:string:boolean:string:string:bigint:string:bigint:boolean:boolean:bigint:boolean:string:double:bigint:string:binary:int:bigint:bigint #### A masked pattern was here #### hive.kafka.max.retries 6 @@ -1536,15 +1521,9 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.wiki_kafka_avro_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct wiki_kafka_avro_table { bool isrobot, string channel, string timestamp, string flags, bool isunpatrolled, string page, string diffurl, i64 added, string comment, i64 commentlength, bool isnew, bool isminor, i64 delta, bool isanonymous, string user, double deltabucket, i64 deleted, string namespace, binary __key, i32 __partition, i64 __offset, i64 __timestamp} serialization.format 1 serialization.lib org.apache.hadoop.hive.kafka.KafkaSerDe storage_handler org.apache.hadoop.hive.kafka.KafkaStorageHandler - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.kafka.KafkaSerDe input format: org.apache.hadoop.hive.kafka.KafkaInputFormat @@ -1621,7 +1600,6 @@ STAGE PLANS: "type" : "string" } ] } - bucket_count -1 bucketing_version 2 column.name.delimiter , columns isrobot,channel,timestamp,flags,isunpatrolled,page,diffurl,added,comment,commentlength,isnew,isminor,delta,isanonymous,user,deltabucket,deleted,namespace,__key,__partition,__offset,__timestamp @@ -1639,15 +1617,9 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.wiki_kafka_avro_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct wiki_kafka_avro_table { bool isrobot, string channel, string timestamp, string flags, bool isunpatrolled, string page, string diffurl, i64 added, string comment, i64 commentlength, bool isnew, bool isminor, i64 delta, bool isanonymous, string user, double deltabucket, i64 deleted, string namespace, binary __key, i32 __partition, i64 __offset, i64 __timestamp} serialization.format 1 serialization.lib org.apache.hadoop.hive.kafka.KafkaSerDe storage_handler org.apache.hadoop.hive.kafka.KafkaStorageHandler - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.kafka.KafkaSerDe name: default.wiki_kafka_avro_table name: default.wiki_kafka_avro_table diff --git a/ql/src/test/results/clientpositive/list_bucket_dml_8.q.out b/ql/src/test/results/clientpositive/list_bucket_dml_8.q.out index 985195850a..bc9bdc7116 100644 --- a/ql/src/test/results/clientpositive/list_bucket_dml_8.q.out +++ b/ql/src/test/results/clientpositive/list_bucket_dml_8.q.out @@ -63,7 +63,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -73,10 +72,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n2 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n2 TotalFiles: 1 @@ -116,30 +113,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -149,10 +136,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -165,30 +150,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -198,10 +173,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -257,7 +230,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -267,10 +239,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n2 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n2 @@ -542,30 +512,20 @@ STAGE PLANS: ds 2008-04-08 hr a1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part_n2 - numFiles 2 - numRows 16 partition_columns ds/hr partition_columns.types string:string - rawDataSize 136 - serialization.ddl struct list_bucketing_dynamic_part_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 310 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -575,10 +535,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n2 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n2 name: default.list_bucketing_dynamic_part_n2 @@ -591,29 +549,20 @@ STAGE PLANS: ds 2008-04-08 hr b1 properties: - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part_n2 - numFiles 3 - numRows 984 partition_columns ds/hr partition_columns.types string:string - rawDataSize 9488 - serialization.ddl struct list_bucketing_dynamic_part_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 10586 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -623,10 +572,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n2 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n2 name: default.list_bucketing_dynamic_part_n2 diff --git a/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out b/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out index 81859fe57b..0e8f3af250 100644 --- a/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out +++ b/ql/src/test/results/clientpositive/llap/acid_bucket_pruning.q.out @@ -97,33 +97,24 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true"}} bucket_count 16 bucket_field_name a bucketing_version 2 column.name.delimiter , columns a - columns.comments columns.types int #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.acidtbldefault - numFiles 17 - numRows 9174 - rawDataSize 0 - serialization.ddl struct acidtbldefault { i32 a} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 34811 transactional true transactional_properties default -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true"}} bucket_count 16 bucket_field_name a bucketing_version 2 @@ -134,16 +125,10 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.acidtbldefault - numFiles 17 - numRows 9174 - rawDataSize 0 - serialization.ddl struct acidtbldefault { i32 a} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 34811 transactional true transactional_properties default -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.acidtbldefault name: default.acidtbldefault diff --git a/ql/src/test/results/clientpositive/llap/acid_nullscan.q.out b/ql/src/test/results/clientpositive/llap/acid_nullscan.q.out index 85d58dddb4..b941555ae8 100644 --- a/ql/src/test/results/clientpositive/llap/acid_nullscan.q.out +++ b/ql/src/test/results/clientpositive/llap/acid_nullscan.q.out @@ -89,32 +89,23 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} bucket_count 2 bucket_field_name a bucketing_version 2 column.name.delimiter , columns a,b - columns.comments columns.types int:string #### A masked pattern was here #### name default.acid_vectorized_n1 - numFiles 3 - numRows 11 - rawDataSize 0 - serialization.ddl struct acid_vectorized_n1 { i32 a, string b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 2583 transactional true transactional_properties default -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} bucket_count 2 bucket_field_name a bucketing_version 2 @@ -124,16 +115,10 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.acid_vectorized_n1 - numFiles 3 - numRows 11 - rawDataSize 0 - serialization.ddl struct acid_vectorized_n1 { i32 a, string b} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 2583 transactional true transactional_properties default -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.acid_vectorized_n1 name: default.acid_vectorized_n1 diff --git a/ql/src/test/results/clientpositive/llap/alter_partition_coltype.q.out b/ql/src/test/results/clientpositive/llap/alter_partition_coltype.q.out index afffbd2862..3567e5c790 100644 --- a/ql/src/test/results/clientpositive/llap/alter_partition_coltype.q.out +++ b/ql/src/test/results/clientpositive/llap/alter_partition_coltype.q.out @@ -263,30 +263,20 @@ STAGE PLANS: dt 100 ts 3.0 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.alter_coltype - numFiles 1 - numRows 25 partition_columns dt/ts partition_columns.types string:double - rawDataSize 191 - serialization.ddl struct alter_coltype { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -296,10 +286,8 @@ STAGE PLANS: name default.alter_coltype partition_columns dt/ts partition_columns.types string:double - serialization.ddl struct alter_coltype { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.alter_coltype name: default.alter_coltype @@ -310,30 +298,20 @@ STAGE PLANS: dt 100 ts 6.30 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.alter_coltype - numFiles 1 - numRows 25 partition_columns dt/ts partition_columns.types string:double - rawDataSize 191 - serialization.ddl struct alter_coltype { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -343,10 +321,8 @@ STAGE PLANS: name default.alter_coltype partition_columns dt/ts partition_columns.types string:double - serialization.ddl struct alter_coltype { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.alter_coltype name: default.alter_coltype @@ -507,30 +483,20 @@ STAGE PLANS: partcol1 1 partcol2 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"intcol":"true"}} - bucket_count -1 column.name.delimiter , columns intcol - columns.comments columns.types string #### A masked pattern was here #### name pt.alterdynamic_part_table - numFiles 2 - numRows 2 partition_columns partcol1/partcol2 partition_columns.types int:string - rawDataSize 3 - serialization.ddl struct alterdynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns intcol @@ -540,10 +506,8 @@ STAGE PLANS: name pt.alterdynamic_part_table partition_columns partcol1/partcol2 partition_columns.types int:string - serialization.ddl struct alterdynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: pt.alterdynamic_part_table name: pt.alterdynamic_part_table @@ -585,30 +549,20 @@ STAGE PLANS: partcol1 2 partcol2 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"intcol":"true"}} - bucket_count -1 column.name.delimiter , columns intcol - columns.comments columns.types string #### A masked pattern was here #### name pt.alterdynamic_part_table - numFiles 1 - numRows 1 partition_columns partcol1/partcol2 partition_columns.types int:string - rawDataSize 1 - serialization.ddl struct alterdynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns intcol @@ -618,10 +572,8 @@ STAGE PLANS: name pt.alterdynamic_part_table partition_columns partcol1/partcol2 partition_columns.types int:string - serialization.ddl struct alterdynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: pt.alterdynamic_part_table name: pt.alterdynamic_part_table diff --git a/ql/src/test/results/clientpositive/llap/analyze_table_null_partition.q.out b/ql/src/test/results/clientpositive/llap/analyze_table_null_partition.q.out index 3206980bd1..da06a7d599 100644 --- a/ql/src/test/results/clientpositive/llap/analyze_table_null_partition.q.out +++ b/ql/src/test/results/clientpositive/llap/analyze_table_null_partition.q.out @@ -115,30 +115,20 @@ STAGE PLANS: partition values: age 15 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"name":"true"}} - bucket_count -1 column.name.delimiter , columns name - columns.comments columns.types string #### A masked pattern was here #### name default.test2_n6 - numFiles 1 - numRows 1 partition_columns age partition_columns.types int - rawDataSize 3 - serialization.ddl struct test2_n6 { string name} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 4 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns name @@ -148,10 +138,8 @@ STAGE PLANS: name default.test2_n6 partition_columns age partition_columns.types int - serialization.ddl struct test2_n6 { string name} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n6 name: default.test2_n6 @@ -161,30 +149,20 @@ STAGE PLANS: partition values: age 30 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"name":"true"}} - bucket_count -1 column.name.delimiter , columns name - columns.comments columns.types string #### A masked pattern was here #### name default.test2_n6 - numFiles 1 - numRows 1 partition_columns age partition_columns.types int - rawDataSize 0 - serialization.ddl struct test2_n6 { string name} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns name @@ -194,10 +172,8 @@ STAGE PLANS: name default.test2_n6 partition_columns age partition_columns.types int - serialization.ddl struct test2_n6 { string name} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n6 name: default.test2_n6 @@ -207,30 +183,20 @@ STAGE PLANS: partition values: age 40 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"name":"true"}} - bucket_count -1 column.name.delimiter , columns name - columns.comments columns.types string #### A masked pattern was here #### name default.test2_n6 - numFiles 1 - numRows 1 partition_columns age partition_columns.types int - rawDataSize 4 - serialization.ddl struct test2_n6 { string name} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns name @@ -240,10 +206,8 @@ STAGE PLANS: name default.test2_n6 partition_columns age partition_columns.types int - serialization.ddl struct test2_n6 { string name} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n6 name: default.test2_n6 @@ -253,30 +217,20 @@ STAGE PLANS: partition values: age __HIVE_DEFAULT_PARTITION__ properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"name":"true"}} - bucket_count -1 column.name.delimiter , columns name - columns.comments columns.types string #### A masked pattern was here #### name default.test2_n6 - numFiles 1 - numRows 2 partition_columns age partition_columns.types int - rawDataSize 4 - serialization.ddl struct test2_n6 { string name} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 6 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns name @@ -286,10 +240,8 @@ STAGE PLANS: name default.test2_n6 partition_columns age partition_columns.types int - serialization.ddl struct test2_n6 { string name} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n6 name: default.test2_n6 diff --git a/ql/src/test/results/clientpositive/llap/autoColumnStats_5a.q.out b/ql/src/test/results/clientpositive/llap/autoColumnStats_5a.q.out index 24b40f0632..e92048ccb3 100644 --- a/ql/src/test/results/clientpositive/llap/autoColumnStats_5a.q.out +++ b/ql/src/test/results/clientpositive/llap/autoColumnStats_5a.q.out @@ -62,7 +62,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -72,10 +71,8 @@ STAGE PLANS: name default.partitioned1 partition_columns part partition_columns.types int - serialization.ddl struct partitioned1 { i32 a, string b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.partitioned1 TotalFiles: 1 @@ -118,11 +115,9 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns - columns.comments columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -130,7 +125,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns @@ -138,7 +132,6 @@ STAGE PLANS: columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -199,7 +192,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -209,10 +201,8 @@ STAGE PLANS: name default.partitioned1 partition_columns part partition_columns.types int - serialization.ddl struct partitioned1 { i32 a, string b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.partitioned1 @@ -331,7 +321,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -341,10 +330,8 @@ STAGE PLANS: name default.partitioned1 partition_columns part partition_columns.types int - serialization.ddl struct partitioned1 { i32 a, string b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.partitioned1 TotalFiles: 1 @@ -387,11 +374,9 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns - columns.comments columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -399,7 +384,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns @@ -407,7 +391,6 @@ STAGE PLANS: columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -468,7 +451,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -478,10 +460,8 @@ STAGE PLANS: name default.partitioned1 partition_columns part partition_columns.types int - serialization.ddl struct partitioned1 { i32 a, string b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.partitioned1 diff --git a/ql/src/test/results/clientpositive/llap/autoColumnStats_8.q.out b/ql/src/test/results/clientpositive/llap/autoColumnStats_8.q.out index 3bb474c927..48d1d92078 100644 --- a/ql/src/test/results/clientpositive/llap/autoColumnStats_8.q.out +++ b/ql/src/test/results/clientpositive/llap/autoColumnStats_8.q.out @@ -98,7 +98,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -108,10 +107,8 @@ STAGE PLANS: name default.nzhang_part8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct nzhang_part8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part8 TotalFiles: 1 @@ -160,7 +157,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -170,10 +166,8 @@ STAGE PLANS: name default.nzhang_part8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct nzhang_part8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part8 TotalFiles: 1 @@ -215,30 +209,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -248,10 +232,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -264,30 +246,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -297,10 +269,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -313,30 +283,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -346,10 +306,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -362,30 +320,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -395,10 +343,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -499,7 +445,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -509,10 +454,8 @@ STAGE PLANS: name default.nzhang_part8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct nzhang_part8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part8 @@ -533,7 +476,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -543,10 +485,8 @@ STAGE PLANS: name default.nzhang_part8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct nzhang_part8 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part8 diff --git a/ql/src/test/results/clientpositive/llap/auto_join_reordering_values.q.out b/ql/src/test/results/clientpositive/llap/auto_join_reordering_values.q.out index e87dc546a0..18f394b971 100644 --- a/ql/src/test/results/clientpositive/llap/auto_join_reordering_values.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_join_reordering_values.q.out @@ -155,30 +155,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns dealid,date,time,cityid,userid - columns.comments columns.types int:string:string:int:int #### A masked pattern was here #### name default.orderpayment_small - numFiles 1 - numRows 1 - rawDataSize 36 - serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 37 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns dealid,date,time,cityid,userid @@ -186,14 +176,8 @@ STAGE PLANS: columns.types int:string:string:int:int #### A masked pattern was here #### name default.orderpayment_small - numFiles 1 - numRows 1 - rawDataSize 36 - serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 37 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.orderpayment_small name: default.orderpayment_small @@ -235,30 +219,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns dealid,date,time,cityid,userid - columns.comments columns.types int:string:string:int:int #### A masked pattern was here #### name default.orderpayment_small - numFiles 1 - numRows 1 - rawDataSize 36 - serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 37 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns dealid,date,time,cityid,userid @@ -266,14 +240,8 @@ STAGE PLANS: columns.types int:string:string:int:int #### A masked pattern was here #### name default.orderpayment_small - numFiles 1 - numRows 1 - rawDataSize 36 - serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 37 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.orderpayment_small name: default.orderpayment_small @@ -315,30 +283,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns dealid,date,time,cityid,userid - columns.comments columns.types int:string:string:int:int #### A masked pattern was here #### name default.orderpayment_small - numFiles 1 - numRows 1 - rawDataSize 36 - serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 37 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns dealid,date,time,cityid,userid @@ -346,14 +304,8 @@ STAGE PLANS: columns.types int:string:string:int:int #### A masked pattern was here #### name default.orderpayment_small - numFiles 1 - numRows 1 - rawDataSize 36 - serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 37 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.orderpayment_small name: default.orderpayment_small @@ -395,30 +347,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns dealid,date,time,cityid,userid - columns.comments columns.types int:string:string:int:int #### A masked pattern was here #### name default.orderpayment_small - numFiles 1 - numRows 1 - rawDataSize 36 - serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 37 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cityid":"true","date":"true","dealid":"true","time":"true","userid":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns dealid,date,time,cityid,userid @@ -426,14 +368,8 @@ STAGE PLANS: columns.types int:string:string:int:int #### A masked pattern was here #### name default.orderpayment_small - numFiles 1 - numRows 1 - rawDataSize 36 - serialization.ddl struct orderpayment_small { i32 dealid, string date, string time, i32 cityid, i32 userid} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 37 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.orderpayment_small name: default.orderpayment_small @@ -475,30 +411,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"userid":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns userid - columns.comments columns.types int #### A masked pattern was here #### name default.user_small - numFiles 1 - numRows 100 - rawDataSize 288 - serialization.ddl struct user_small { i32 userid} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 388 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"userid":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns userid @@ -506,14 +432,8 @@ STAGE PLANS: columns.types int #### A masked pattern was here #### name default.user_small - numFiles 1 - numRows 100 - rawDataSize 288 - serialization.ddl struct user_small { i32 userid} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 388 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.user_small name: default.user_small diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_1.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_1.q.out index 096e225844..7566b3476f 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_1.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_1.q.out @@ -168,20 +168,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n1 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -199,10 +192,8 @@ STAGE PLANS: name default.bucket_small_n1 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n1 name: default.bucket_small_n1 @@ -260,20 +251,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n1 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -291,10 +275,8 @@ STAGE PLANS: name default.bucket_big_n1 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n1 name: default.bucket_big_n1 @@ -309,20 +291,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n1 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -340,10 +315,8 @@ STAGE PLANS: name default.bucket_big_n1 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n1 name: default.bucket_big_n1 @@ -472,20 +445,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n1 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -503,10 +469,8 @@ STAGE PLANS: name default.bucket_small_n1 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n1 name: default.bucket_small_n1 @@ -564,20 +528,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n1 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -595,10 +552,8 @@ STAGE PLANS: name default.bucket_big_n1 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n1 name: default.bucket_big_n1 @@ -613,20 +568,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n1 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -644,10 +592,8 @@ STAGE PLANS: name default.bucket_big_n1 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n1 name: default.bucket_big_n1 @@ -776,20 +722,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n1 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -807,10 +746,8 @@ STAGE PLANS: name default.bucket_small_n1 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n1 name: default.bucket_small_n1 @@ -868,20 +805,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n1 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -899,10 +829,8 @@ STAGE PLANS: name default.bucket_big_n1 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n1 name: default.bucket_big_n1 @@ -917,20 +845,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n1 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -948,10 +869,8 @@ STAGE PLANS: name default.bucket_big_n1 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n1 name: default.bucket_big_n1 diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_11.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_11.q.out index 4d428a6040..9902ff3ada 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_11.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_11.q.out @@ -179,20 +179,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n11 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -210,10 +203,8 @@ STAGE PLANS: name default.bucket_small_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n11 name: default.bucket_small_n11 @@ -261,20 +252,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n11 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -292,10 +276,8 @@ STAGE PLANS: name default.bucket_big_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n11 name: default.bucket_big_n11 @@ -311,20 +293,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n11 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -342,10 +317,8 @@ STAGE PLANS: name default.bucket_big_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n11 name: default.bucket_big_n11 @@ -501,20 +474,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n11 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -532,10 +498,8 @@ STAGE PLANS: name default.bucket_small_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n11 name: default.bucket_small_n11 @@ -593,20 +557,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n11 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -624,10 +581,8 @@ STAGE PLANS: name default.bucket_big_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n11 name: default.bucket_big_n11 @@ -642,20 +597,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n11 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -673,10 +621,8 @@ STAGE PLANS: name default.bucket_big_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n11 name: default.bucket_big_n11 @@ -794,20 +740,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n11 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -825,10 +764,8 @@ STAGE PLANS: name default.bucket_small_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n11 name: default.bucket_small_n11 @@ -882,20 +819,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n11 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -913,10 +843,8 @@ STAGE PLANS: name default.bucket_big_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n11 name: default.bucket_big_n11 @@ -931,20 +859,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n11 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -962,10 +883,8 @@ STAGE PLANS: name default.bucket_big_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n11 name: default.bucket_big_n11 @@ -1084,20 +1003,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n11 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1115,10 +1027,8 @@ STAGE PLANS: name default.bucket_small_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n11 name: default.bucket_small_n11 @@ -1168,20 +1078,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n11 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1199,10 +1102,8 @@ STAGE PLANS: name default.bucket_big_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n11 name: default.bucket_big_n11 @@ -1217,20 +1118,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n11 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1248,10 +1142,8 @@ STAGE PLANS: name default.bucket_big_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n11 name: default.bucket_big_n11 @@ -1296,20 +1188,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n11 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1327,10 +1212,8 @@ STAGE PLANS: name default.bucket_big_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n11 name: default.bucket_big_n11 @@ -1346,20 +1229,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n11 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1377,10 +1253,8 @@ STAGE PLANS: name default.bucket_big_n11 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n11 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n11 name: default.bucket_big_n11 diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_12.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_12.q.out index cf73803c06..d0d01fe93e 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_12.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_12.q.out @@ -223,20 +223,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_medium - numFiles 3 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_medium { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 170 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -254,10 +247,8 @@ STAGE PLANS: name default.bucket_medium partition_columns ds partition_columns.types string - serialization.ddl struct bucket_medium { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_medium name: default.bucket_medium @@ -293,20 +284,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_medium - numFiles 3 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_medium { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 170 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -324,10 +308,8 @@ STAGE PLANS: name default.bucket_medium partition_columns ds partition_columns.types string - serialization.ddl struct bucket_medium { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_medium name: default.bucket_medium @@ -382,20 +364,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n15 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n15 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -413,10 +388,8 @@ STAGE PLANS: name default.bucket_big_n15 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n15 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n15 name: default.bucket_big_n15 @@ -431,20 +404,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n15 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n15 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -462,10 +428,8 @@ STAGE PLANS: name default.bucket_big_n15 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n15 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n15 name: default.bucket_big_n15 @@ -514,20 +478,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n15 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n15 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -545,10 +502,8 @@ STAGE PLANS: name default.bucket_small_n15 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n15 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n15 name: default.bucket_small_n15 diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out index 60cfb52549..fbdf2d6f17 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_2.q.out @@ -150,20 +150,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n3 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -181,10 +174,8 @@ STAGE PLANS: name default.bucket_small_n3 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n3 name: default.bucket_small_n3 @@ -242,20 +233,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n3 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -273,10 +257,8 @@ STAGE PLANS: name default.bucket_big_n3 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n3 name: default.bucket_big_n3 @@ -291,20 +273,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n3 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -322,10 +297,8 @@ STAGE PLANS: name default.bucket_big_n3 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n3 name: default.bucket_big_n3 @@ -454,20 +427,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n3 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -485,10 +451,8 @@ STAGE PLANS: name default.bucket_small_n3 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n3 name: default.bucket_small_n3 @@ -546,20 +510,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n3 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -577,10 +534,8 @@ STAGE PLANS: name default.bucket_big_n3 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n3 name: default.bucket_big_n3 @@ -595,20 +550,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n3 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -626,10 +574,8 @@ STAGE PLANS: name default.bucket_big_n3 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n3 name: default.bucket_big_n3 diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out index a9b2f03838..be00f1d230 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_3.q.out @@ -150,20 +150,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n9 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -181,10 +174,8 @@ STAGE PLANS: name default.bucket_small_n9 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n9 name: default.bucket_small_n9 @@ -199,20 +190,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n9 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -230,10 +214,8 @@ STAGE PLANS: name default.bucket_small_n9 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n9 name: default.bucket_small_n9 @@ -292,20 +274,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n9 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -323,10 +298,8 @@ STAGE PLANS: name default.bucket_big_n9 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n9 name: default.bucket_big_n9 @@ -454,20 +427,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n9 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -485,10 +451,8 @@ STAGE PLANS: name default.bucket_small_n9 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n9 name: default.bucket_small_n9 @@ -503,20 +467,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n9 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -534,10 +491,8 @@ STAGE PLANS: name default.bucket_small_n9 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n9 name: default.bucket_small_n9 @@ -596,20 +551,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n9 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -627,10 +575,8 @@ STAGE PLANS: name default.bucket_big_n9 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n9 name: default.bucket_big_n9 @@ -758,20 +704,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n9 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -789,10 +728,8 @@ STAGE PLANS: name default.bucket_small_n9 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n9 name: default.bucket_small_n9 @@ -807,20 +744,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n9 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -838,10 +768,8 @@ STAGE PLANS: name default.bucket_small_n9 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n9 name: default.bucket_small_n9 @@ -900,20 +828,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n9 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -931,10 +852,8 @@ STAGE PLANS: name default.bucket_big_n9 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n9 name: default.bucket_big_n9 diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_4.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_4.q.out index 6ef466fa20..20937079ef 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_4.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_4.q.out @@ -166,20 +166,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n12 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -197,10 +190,8 @@ STAGE PLANS: name default.bucket_small_n12 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n12 name: default.bucket_small_n12 @@ -215,20 +206,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n12 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -246,10 +230,8 @@ STAGE PLANS: name default.bucket_small_n12 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n12 name: default.bucket_small_n12 @@ -308,20 +290,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n12 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -339,10 +314,8 @@ STAGE PLANS: name default.bucket_big_n12 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n12 name: default.bucket_big_n12 @@ -470,20 +443,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n12 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -501,10 +467,8 @@ STAGE PLANS: name default.bucket_small_n12 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n12 name: default.bucket_small_n12 @@ -519,20 +483,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n12 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -550,10 +507,8 @@ STAGE PLANS: name default.bucket_small_n12 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n12 name: default.bucket_small_n12 @@ -612,20 +567,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n12 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -643,10 +591,8 @@ STAGE PLANS: name default.bucket_big_n12 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n12 name: default.bucket_big_n12 @@ -774,20 +720,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n12 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -805,10 +744,8 @@ STAGE PLANS: name default.bucket_small_n12 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n12 name: default.bucket_small_n12 @@ -823,20 +760,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n12 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -854,10 +784,8 @@ STAGE PLANS: name default.bucket_small_n12 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n12 name: default.bucket_small_n12 @@ -916,20 +844,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n12 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -947,10 +868,8 @@ STAGE PLANS: name default.bucket_big_n12 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n12 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n12 name: default.bucket_big_n12 diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_5.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_5.q.out index f9c09d5c41..0ffe70ed20 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_5.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_5.q.out @@ -125,18 +125,11 @@ STAGE PLANS: bucketing_version 1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n0 - numFiles 4 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_small_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -152,14 +145,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small_n0 - numFiles 4 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_small_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n0 name: default.bucket_small_n0 @@ -217,18 +204,11 @@ STAGE PLANS: bucketing_version 1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n0 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_big_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -244,14 +224,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big_n0 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_big_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n0 name: default.bucket_big_n0 @@ -367,18 +341,11 @@ STAGE PLANS: bucketing_version 1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n0 - numFiles 4 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_small_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -394,14 +361,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small_n0 - numFiles 4 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_small_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n0 name: default.bucket_small_n0 @@ -459,18 +420,11 @@ STAGE PLANS: bucketing_version 1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n0 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_big_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -486,14 +440,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big_n0 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_big_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n0 name: default.bucket_big_n0 @@ -609,18 +557,11 @@ STAGE PLANS: bucketing_version 1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n0 - numFiles 4 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_small_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -636,14 +577,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_small_n0 - numFiles 4 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_small_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n0 name: default.bucket_small_n0 @@ -701,18 +636,11 @@ STAGE PLANS: bucketing_version 1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n0 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_big_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -728,14 +656,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.bucket_big_n0 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_big_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n0 name: default.bucket_big_n0 diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_7.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_7.q.out index a20ac713bd..243435e366 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_7.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_7.q.out @@ -185,20 +185,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n6 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -216,10 +209,8 @@ STAGE PLANS: name default.bucket_small_n6 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n6 name: default.bucket_small_n6 @@ -234,20 +225,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n6 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -265,10 +249,8 @@ STAGE PLANS: name default.bucket_small_n6 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n6 name: default.bucket_small_n6 @@ -327,20 +309,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n6 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -358,10 +333,8 @@ STAGE PLANS: name default.bucket_big_n6 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n6 name: default.bucket_big_n6 @@ -376,20 +349,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n6 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -407,10 +373,8 @@ STAGE PLANS: name default.bucket_big_n6 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n6 name: default.bucket_big_n6 @@ -543,20 +507,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n6 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -574,10 +531,8 @@ STAGE PLANS: name default.bucket_small_n6 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n6 name: default.bucket_small_n6 @@ -592,20 +547,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n6 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -623,10 +571,8 @@ STAGE PLANS: name default.bucket_small_n6 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n6 name: default.bucket_small_n6 @@ -685,20 +631,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n6 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -716,10 +655,8 @@ STAGE PLANS: name default.bucket_big_n6 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n6 name: default.bucket_big_n6 @@ -734,20 +671,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n6 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -765,10 +695,8 @@ STAGE PLANS: name default.bucket_big_n6 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n6 name: default.bucket_big_n6 @@ -901,20 +829,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n6 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -932,10 +853,8 @@ STAGE PLANS: name default.bucket_small_n6 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n6 name: default.bucket_small_n6 @@ -950,20 +869,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n6 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 226 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -981,10 +893,8 @@ STAGE PLANS: name default.bucket_small_n6 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n6 name: default.bucket_small_n6 @@ -1043,20 +953,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n6 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1074,10 +977,8 @@ STAGE PLANS: name default.bucket_big_n6 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n6 name: default.bucket_big_n6 @@ -1092,20 +993,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n6 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1123,10 +1017,8 @@ STAGE PLANS: name default.bucket_big_n6 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n6 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n6 name: default.bucket_big_n6 diff --git a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_8.q.out b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_8.q.out index 9b5a8ef36f..efdb245f77 100644 --- a/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_8.q.out +++ b/ql/src/test/results/clientpositive/llap/auto_sortmerge_join_8.q.out @@ -185,20 +185,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n5 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -216,10 +209,8 @@ STAGE PLANS: name default.bucket_small_n5 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n5 name: default.bucket_small_n5 @@ -234,20 +225,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n5 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -265,10 +249,8 @@ STAGE PLANS: name default.bucket_small_n5 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n5 name: default.bucket_small_n5 @@ -327,20 +309,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n5 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -358,10 +333,8 @@ STAGE PLANS: name default.bucket_big_n5 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n5 name: default.bucket_big_n5 @@ -376,20 +349,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n5 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -407,10 +373,8 @@ STAGE PLANS: name default.bucket_big_n5 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n5 name: default.bucket_big_n5 @@ -543,20 +507,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n5 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -574,10 +531,8 @@ STAGE PLANS: name default.bucket_small_n5 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n5 name: default.bucket_small_n5 @@ -592,20 +547,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n5 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -623,10 +571,8 @@ STAGE PLANS: name default.bucket_small_n5 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n5 name: default.bucket_small_n5 @@ -685,20 +631,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n5 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -716,10 +655,8 @@ STAGE PLANS: name default.bucket_big_n5 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n5 name: default.bucket_big_n5 @@ -734,20 +671,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n5 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -765,10 +695,8 @@ STAGE PLANS: name default.bucket_big_n5 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n5 name: default.bucket_big_n5 @@ -901,20 +829,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n5 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -932,10 +853,8 @@ STAGE PLANS: name default.bucket_small_n5 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n5 name: default.bucket_small_n5 @@ -950,20 +869,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_small_n5 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_small_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 114 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -981,10 +893,8 @@ STAGE PLANS: name default.bucket_small_n5 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_small_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_small_n5 name: default.bucket_small_n5 @@ -1043,20 +953,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n5 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1074,10 +977,8 @@ STAGE PLANS: name default.bucket_big_n5 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n5 name: default.bucket_big_n5 @@ -1092,20 +993,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.bucket_big_n5 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct bucket_big_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1123,10 +1017,8 @@ STAGE PLANS: name default.bucket_big_n5 partition_columns ds partition_columns.types string - serialization.ddl struct bucket_big_n5 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_big_n5 name: default.bucket_big_n5 diff --git a/ql/src/test/results/clientpositive/llap/bucket1.q.out b/ql/src/test/results/clientpositive/llap/bucket1.q.out index 3ec523f461..706d391ea9 100644 --- a/ql/src/test/results/clientpositive/llap/bucket1.q.out +++ b/ql/src/test/results/clientpositive/llap/bucket1.q.out @@ -67,30 +67,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -98,14 +88,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -131,7 +115,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 100 bucket_field_name key bucketing_version 2 @@ -141,14 +124,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.bucket1_1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket1_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket1_1 TotalFiles: 1 @@ -219,7 +196,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 100 bucket_field_name key bucketing_version 2 @@ -229,14 +205,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.bucket1_1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket1_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket1_1 diff --git a/ql/src/test/results/clientpositive/llap/bucket2.q.out b/ql/src/test/results/clientpositive/llap/bucket2.q.out index 1c1db2f114..d40bd0107b 100644 --- a/ql/src/test/results/clientpositive/llap/bucket2.q.out +++ b/ql/src/test/results/clientpositive/llap/bucket2.q.out @@ -66,30 +66,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -97,14 +87,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -130,7 +114,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -140,14 +123,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.bucket2_1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket2_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket2_1 TotalFiles: 2 @@ -199,7 +176,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -209,14 +185,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.bucket2_1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket2_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket2_1 diff --git a/ql/src/test/results/clientpositive/llap/bucket3.q.out b/ql/src/test/results/clientpositive/llap/bucket3.q.out index 3b303bd76b..094286123e 100644 --- a/ql/src/test/results/clientpositive/llap/bucket3.q.out +++ b/ql/src/test/results/clientpositive/llap/bucket3.q.out @@ -67,30 +67,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -98,14 +88,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -143,10 +127,8 @@ STAGE PLANS: name default.bucket3_1 partition_columns ds partition_columns.types string - serialization.ddl struct bucket3_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket3_1 TotalFiles: 2 @@ -238,10 +220,8 @@ STAGE PLANS: name default.bucket3_1 partition_columns ds partition_columns.types string - serialization.ddl struct bucket3_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket3_1 diff --git a/ql/src/test/results/clientpositive/llap/bucket4.q.out b/ql/src/test/results/clientpositive/llap/bucket4.q.out index b3831e3169..7246806692 100644 --- a/ql/src/test/results/clientpositive/llap/bucket4.q.out +++ b/ql/src/test/results/clientpositive/llap/bucket4.q.out @@ -66,30 +66,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -97,14 +87,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -130,7 +114,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -141,14 +124,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.bucket4_1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket4_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket4_1 TotalFiles: 2 @@ -200,7 +177,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -211,14 +187,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.bucket4_1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket4_1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket4_1 diff --git a/ql/src/test/results/clientpositive/llap/bucket5.q.out b/ql/src/test/results/clientpositive/llap/bucket5.q.out index acb968d90f..7127f021c4 100644 --- a/ql/src/test/results/clientpositive/llap/bucket5.q.out +++ b/ql/src/test/results/clientpositive/llap/bucket5.q.out @@ -101,31 +101,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -134,14 +124,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -167,7 +151,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -179,14 +162,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.bucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketed_table { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketed_table TotalFiles: 1 @@ -264,8 +241,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -274,14 +249,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.unbucketed_table TotalFiles: 1 @@ -352,7 +321,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -364,14 +332,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.bucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketed_table { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketed_table @@ -394,8 +356,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -404,14 +364,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.unbucketed_table @@ -453,8 +407,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -463,14 +415,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.unbucketed_table TotalFiles: 1 @@ -485,8 +431,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -495,21 +439,13 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -518,14 +454,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.unbucketed_table name: default.unbucketed_table @@ -550,8 +480,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -560,14 +488,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.unbucketed_table TotalFiles: 1 @@ -582,8 +504,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -592,21 +512,13 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -615,14 +527,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.unbucketed_table - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct unbucketed_table { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.unbucketed_table name: default.unbucketed_table diff --git a/ql/src/test/results/clientpositive/llap/bucket_many.q.out b/ql/src/test/results/clientpositive/llap/bucket_many.q.out index b478b64c3a..75f3452e17 100644 --- a/ql/src/test/results/clientpositive/llap/bucket_many.q.out +++ b/ql/src/test/results/clientpositive/llap/bucket_many.q.out @@ -67,30 +67,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -98,14 +88,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -131,7 +115,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 256 bucket_field_name key bucketing_version 2 @@ -141,14 +124,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.bucket_many - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_many { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_many TotalFiles: 256 @@ -219,7 +196,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 256 bucket_field_name key bucketing_version 2 @@ -229,14 +205,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.bucket_many - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_many { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_many diff --git a/ql/src/test/results/clientpositive/llap/bucket_map_join_tez2.q.out b/ql/src/test/results/clientpositive/llap/bucket_map_join_tez2.q.out index 414b143ce5..54082b0ae3 100644 --- a/ql/src/test/results/clientpositive/llap/bucket_map_join_tez2.q.out +++ b/ql/src/test/results/clientpositive/llap/bucket_map_join_tez2.q.out @@ -2524,30 +2524,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} bucket_count -1 bucketing_version 2 column.name.delimiter , columns join_col,filter_col - columns.comments columns.types string:string #### A masked pattern was here #### name default.my_dim - numFiles 1 - numRows 4 - rawDataSize 692 - serialization.ddl struct my_dim { string join_col, string filter_col} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 340 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns join_col,filter_col @@ -2555,14 +2545,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.my_dim - numFiles 1 - numRows 4 - rawDataSize 692 - serialization.ddl struct my_dim { string join_col, string filter_col} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 340 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.my_dim name: default.my_dim diff --git a/ql/src/test/results/clientpositive/llap/bucket_num_reducers.q.out b/ql/src/test/results/clientpositive/llap/bucket_num_reducers.q.out index ccc177bdae..dc05fb232d 100644 --- a/ql/src/test/results/clientpositive/llap/bucket_num_reducers.q.out +++ b/ql/src/test/results/clientpositive/llap/bucket_num_reducers.q.out @@ -64,30 +64,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -95,14 +85,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -128,7 +112,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 50 bucket_field_name key bucketing_version 2 @@ -138,14 +121,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.bucket_nr - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_nr { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_nr TotalFiles: 50 @@ -164,7 +141,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 50 bucket_field_name key bucketing_version 2 @@ -174,14 +150,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.bucket_nr - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket_nr { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket_nr diff --git a/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out b/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out index 5ea3d6f14e..716d60a822 100644 --- a/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out +++ b/ql/src/test/results/clientpositive/llap/bucket_num_reducers2.q.out @@ -64,30 +64,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -95,14 +85,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -128,7 +112,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 3 bucket_field_name key bucketing_version 2 @@ -138,14 +121,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.test_table_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct test_table_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table_n4 TotalFiles: 3 @@ -197,7 +174,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 3 bucket_field_name key bucketing_version 2 @@ -207,14 +183,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.test_table_n4 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct test_table_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table_n4 diff --git a/ql/src/test/results/clientpositive/llap/bucketmapjoin1.q.out b/ql/src/test/results/clientpositive/llap/bucketmapjoin1.q.out index a5b97e1faa..018cb75ef3 100644 --- a/ql/src/test/results/clientpositive/llap/bucketmapjoin1.q.out +++ b/ql/src/test/results/clientpositive/llap/bucketmapjoin1.q.out @@ -502,18 +502,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n1 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -528,14 +521,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n1 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_n1 name: default.srcbucket_mapjoin_n1 @@ -584,20 +571,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n1 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -614,10 +594,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n1 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n1 name: default.srcbucket_mapjoin_part_n1 @@ -652,8 +630,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -661,14 +637,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n0 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n0 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n0 TotalFiles: 1 @@ -739,8 +709,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -748,14 +716,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n0 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n0 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n0 @@ -954,18 +916,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n1 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -980,14 +935,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n1 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_n1 name: default.srcbucket_mapjoin_n1 @@ -1036,20 +985,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n1 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1066,10 +1008,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n1 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n1 name: default.srcbucket_mapjoin_part_n1 @@ -1104,8 +1044,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -1113,14 +1051,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n0 - numFiles 1 - numRows 464 - rawDataSize 8519 - serialization.ddl struct bucketmapjoin_tmp_result_n0 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 8983 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n0 TotalFiles: 1 @@ -1191,8 +1123,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -1200,14 +1130,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n0 - numFiles 1 - numRows 464 - rawDataSize 8519 - serialization.ddl struct bucketmapjoin_tmp_result_n0 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 8983 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n0 diff --git a/ql/src/test/results/clientpositive/llap/bucketmapjoin2.q.out b/ql/src/test/results/clientpositive/llap/bucketmapjoin2.q.out index f5a8dc2dfd..bc9afb00eb 100644 --- a/ql/src/test/results/clientpositive/llap/bucketmapjoin2.q.out +++ b/ql/src/test/results/clientpositive/llap/bucketmapjoin2.q.out @@ -175,20 +175,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n6 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -205,10 +198,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n6 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n6 name: default.srcbucket_mapjoin_part_n6 @@ -257,20 +248,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n5 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -287,10 +271,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n5 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n5 name: default.srcbucket_mapjoin_part_2_n5 @@ -325,8 +307,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -334,14 +314,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n2 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n2 TotalFiles: 1 @@ -412,8 +386,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -421,14 +393,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n2 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n2 @@ -634,20 +600,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n6 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -664,10 +623,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n6 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n6 name: default.srcbucket_mapjoin_part_n6 @@ -716,20 +673,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n5 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -746,10 +696,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n5 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n5 name: default.srcbucket_mapjoin_part_2_n5 @@ -784,8 +732,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -793,14 +739,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n2 - numFiles 1 - numRows 564 - rawDataSize 10503 - serialization.ddl struct bucketmapjoin_tmp_result_n2 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11067 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n2 TotalFiles: 1 @@ -871,8 +811,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -880,14 +818,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n2 - numFiles 1 - numRows 564 - rawDataSize 10503 - serialization.ddl struct bucketmapjoin_tmp_result_n2 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11067 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n2 @@ -1112,20 +1044,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n6 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1142,10 +1067,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n6 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n6 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n6 name: default.srcbucket_mapjoin_part_n6 @@ -1194,20 +1117,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n5 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1224,10 +1140,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n5 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n5 name: default.srcbucket_mapjoin_part_2_n5 @@ -1243,20 +1157,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n5 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1273,10 +1180,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n5 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n5 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n5 name: default.srcbucket_mapjoin_part_2_n5 @@ -1312,8 +1217,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -1321,14 +1224,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n2 - numFiles 1 - numRows 564 - rawDataSize 10503 - serialization.ddl struct bucketmapjoin_tmp_result_n2 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11067 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n2 TotalFiles: 1 @@ -1399,8 +1296,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -1408,14 +1303,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n2 - numFiles 1 - numRows 564 - rawDataSize 10503 - serialization.ddl struct bucketmapjoin_tmp_result_n2 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11067 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n2 diff --git a/ql/src/test/results/clientpositive/llap/bucketmapjoin3.q.out b/ql/src/test/results/clientpositive/llap/bucketmapjoin3.q.out index e11bb747be..c2ec2c208c 100644 --- a/ql/src/test/results/clientpositive/llap/bucketmapjoin3.q.out +++ b/ql/src/test/results/clientpositive/llap/bucketmapjoin3.q.out @@ -199,20 +199,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n11 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n11 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -229,10 +222,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n11 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n11 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n11 name: default.srcbucket_mapjoin_part_2_n11 @@ -281,20 +272,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n13 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n13 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -311,10 +295,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n13 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n13 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n13 name: default.srcbucket_mapjoin_part_n13 @@ -349,8 +331,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -358,14 +338,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n6 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n6 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n6 TotalFiles: 1 @@ -436,8 +410,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -445,14 +417,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n6 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n6 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n6 @@ -658,20 +624,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n11 - numFiles 2 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n11 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3062 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -688,10 +647,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n11 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n11 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n11 name: default.srcbucket_mapjoin_part_2_n11 @@ -740,20 +697,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n13 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n13 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -770,10 +720,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n13 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n13 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n13 name: default.srcbucket_mapjoin_part_n13 @@ -808,8 +756,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -817,14 +763,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n6 - numFiles 1 - numRows 564 - rawDataSize 10503 - serialization.ddl struct bucketmapjoin_tmp_result_n6 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11067 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n6 TotalFiles: 1 @@ -895,8 +835,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -904,14 +842,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n6 - numFiles 1 - numRows 564 - rawDataSize 10503 - serialization.ddl struct bucketmapjoin_tmp_result_n6 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11067 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n6 diff --git a/ql/src/test/results/clientpositive/llap/bucketmapjoin4.q.out b/ql/src/test/results/clientpositive/llap/bucketmapjoin4.q.out index b4d9d3d909..b1976d2d4e 100644 --- a/ql/src/test/results/clientpositive/llap/bucketmapjoin4.q.out +++ b/ql/src/test/results/clientpositive/llap/bucketmapjoin4.q.out @@ -192,18 +192,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n17 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n17 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -218,14 +211,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n17 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n17 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_n17 name: default.srcbucket_mapjoin_n17 @@ -273,18 +260,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n17 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n17 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -299,14 +279,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n17 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n17 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_n17 name: default.srcbucket_mapjoin_n17 @@ -341,8 +315,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -350,14 +322,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n8 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n8 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n8 TotalFiles: 1 @@ -428,8 +394,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -437,14 +401,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n8 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n8 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n8 @@ -631,18 +589,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n17 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n17 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -657,14 +608,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n17 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n17 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_n17 name: default.srcbucket_mapjoin_n17 @@ -712,18 +657,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n17 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n17 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -738,14 +676,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n17 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n17 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_n17 name: default.srcbucket_mapjoin_n17 @@ -780,8 +712,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -789,14 +719,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n8 - numFiles 1 - numRows 464 - rawDataSize 8519 - serialization.ddl struct bucketmapjoin_tmp_result_n8 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 8983 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n8 TotalFiles: 1 @@ -867,8 +791,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -876,14 +798,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n8 - numFiles 1 - numRows 464 - rawDataSize 8519 - serialization.ddl struct bucketmapjoin_tmp_result_n8 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 8983 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n8 diff --git a/ql/src/test/results/clientpositive/llap/bucketmapjoin7.q.out b/ql/src/test/results/clientpositive/llap/bucketmapjoin7.q.out index 5b476edf30..11c263244e 100644 --- a/ql/src/test/results/clientpositive/llap/bucketmapjoin7.q.out +++ b/ql/src/test/results/clientpositive/llap/bucketmapjoin7.q.out @@ -126,20 +126,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n4 - numFiles 2 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_1_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -156,10 +149,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n4 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcbucket_mapjoin_part_1_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n4 name: default.srcbucket_mapjoin_part_1_n4 @@ -205,20 +196,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n9 - numFiles 2 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_2_n9 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -235,10 +219,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n9 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcbucket_mapjoin_part_2_n9 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n9 name: default.srcbucket_mapjoin_part_2_n9 diff --git a/ql/src/test/results/clientpositive/llap/cbo_rp_outer_join_ppr.q.out b/ql/src/test/results/clientpositive/llap/cbo_rp_outer_join_ppr.q.out index c5e10689ea..3d0f5ad2ec 100644 --- a/ql/src/test/results/clientpositive/llap/cbo_rp_outer_join_ppr.q.out +++ b/ql/src/test/results/clientpositive/llap/cbo_rp_outer_join_ppr.q.out @@ -75,30 +75,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -106,14 +96,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -159,30 +143,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -192,10 +166,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -208,30 +180,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -241,10 +203,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -414,30 +374,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -445,14 +395,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -498,30 +442,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -531,10 +465,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -547,30 +479,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -580,10 +502,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/llap/cbo_stats_estimation.q.out b/ql/src/test/results/clientpositive/llap/cbo_stats_estimation.q.out index 03826ba703..9f2731d61a 100644 --- a/ql/src/test/results/clientpositive/llap/cbo_stats_estimation.q.out +++ b/ql/src/test/results/clientpositive/llap/cbo_stats_estimation.q.out @@ -80,24 +80,16 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns claim_rec_id,claim_invoice_num,typ_c - columns.comments columns.types bigint:string:int #### A masked pattern was here #### name default.claims - numFiles 0 - numRows 1154941534 - rawDataSize 1135307527922 - serialization.ddl struct claims { i64 claim_rec_id, string claim_invoice_num, i32 typ_c} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns claim_rec_id,claim_invoice_num,typ_c @@ -105,14 +97,8 @@ STAGE PLANS: columns.types bigint:string:int #### A masked pattern was here #### name default.claims - numFiles 0 - numRows 1154941534 - rawDataSize 1135307527922 - serialization.ddl struct claims { i64 claim_rec_id, string claim_invoice_num, i32 typ_c} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.claims name: default.claims @@ -224,24 +210,16 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns claim_rec_id,claim_invoice_num,typ_c - columns.comments columns.types bigint:string:int #### A masked pattern was here #### name default.claims - numFiles 0 - numRows 1154941534 - rawDataSize 1135307527922 - serialization.ddl struct claims { i64 claim_rec_id, string claim_invoice_num, i32 typ_c} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns claim_rec_id,claim_invoice_num,typ_c @@ -249,14 +227,8 @@ STAGE PLANS: columns.types bigint:string:int #### A masked pattern was here #### name default.claims - numFiles 0 - numRows 1154941534 - rawDataSize 1135307527922 - serialization.ddl struct claims { i64 claim_rec_id, string claim_invoice_num, i32 typ_c} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.claims name: default.claims diff --git a/ql/src/test/results/clientpositive/llap/column_table_stats.q.out b/ql/src/test/results/clientpositive/llap/column_table_stats.q.out index b6d3bb446f..fb4480863a 100644 --- a/ql/src/test/results/clientpositive/llap/column_table_stats.q.out +++ b/ql/src/test/results/clientpositive/llap/column_table_stats.q.out @@ -115,24 +115,16 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.s - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct s { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -140,14 +132,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.s - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct s { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.s name: default.s @@ -393,29 +379,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.spart - numFiles 1 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct spart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -425,10 +402,8 @@ STAGE PLANS: name default.spart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct spart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.spart name: default.spart @@ -441,29 +416,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.spart - numFiles 1 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct spart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -473,10 +439,8 @@ STAGE PLANS: name default.spart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct spart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.spart name: default.spart @@ -822,29 +786,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.spart - numFiles 1 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct spart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -854,10 +809,8 @@ STAGE PLANS: name default.spart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct spart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.spart name: default.spart @@ -870,29 +823,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.spart - numFiles 1 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct spart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -902,10 +846,8 @@ STAGE PLANS: name default.spart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct spart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.spart name: default.spart @@ -1248,29 +1190,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.spart - numFiles 1 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct spart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1280,10 +1213,8 @@ STAGE PLANS: name default.spart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct spart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.spart name: default.spart diff --git a/ql/src/test/results/clientpositive/llap/column_table_stats_orc.q.out b/ql/src/test/results/clientpositive/llap/column_table_stats_orc.q.out index 7ca66229fb..dc80b7ccad 100644 --- a/ql/src/test/results/clientpositive/llap/column_table_stats_orc.q.out +++ b/ql/src/test/results/clientpositive/llap/column_table_stats_orc.q.out @@ -114,30 +114,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.s_n0 - numFiles 1 - numRows 1 - rawDataSize 170 - serialization.ddl struct s_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 285 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -145,14 +135,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.s_n0 - numFiles 1 - numRows 1 - rawDataSize 170 - serialization.ddl struct s_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 285 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.s_n0 name: default.s_n0 @@ -396,26 +380,17 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 #### A masked pattern was here #### name default.spart_n0 - numFiles 1 - numRows 1 partition_columns ds/hr partition_columns.types string:string - rawDataSize 170 - serialization.ddl struct spart_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 285 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -425,10 +400,8 @@ STAGE PLANS: name default.spart_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct spart_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.spart_n0 name: default.spart_n0 @@ -441,26 +414,17 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 #### A masked pattern was here #### name default.spart_n0 - numFiles 1 - numRows 1 partition_columns ds/hr partition_columns.types string:string - rawDataSize 170 - serialization.ddl struct spart_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 285 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -470,10 +434,8 @@ STAGE PLANS: name default.spart_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct spart_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.spart_n0 name: default.spart_n0 @@ -814,26 +776,17 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 #### A masked pattern was here #### name default.spart_n0 - numFiles 1 - numRows 1 partition_columns ds/hr partition_columns.types string:string - rawDataSize 170 - serialization.ddl struct spart_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 285 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -843,10 +796,8 @@ STAGE PLANS: name default.spart_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct spart_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.spart_n0 name: default.spart_n0 diff --git a/ql/src/test/results/clientpositive/llap/columnstats_partlvl.q.out b/ql/src/test/results/clientpositive/llap/columnstats_partlvl.q.out index 4048ee9410..8152a824f1 100644 --- a/ql/src/test/results/clientpositive/llap/columnstats_partlvl.q.out +++ b/ql/src/test/results/clientpositive/llap/columnstats_partlvl.q.out @@ -184,30 +184,21 @@ STAGE PLANS: partition values: employeesalary 2000.0 properties: - bucket_count -1 column.name.delimiter , columns employeeid,employeename - columns.comments columns.types int:string field.delim | #### A masked pattern was here #### name default.employee_part - numFiles 1 - numRows 0 partition_columns employeesalary partition_columns.types double - rawDataSize 0 - serialization.ddl struct employee_part { i32 employeeid, string employeename} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 105 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns employeeid,employeename @@ -218,10 +209,8 @@ STAGE PLANS: name default.employee_part partition_columns employeesalary partition_columns.types double - serialization.ddl struct employee_part { i32 employeeid, string employeename} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.employee_part name: default.employee_part @@ -482,30 +471,21 @@ STAGE PLANS: partition values: employeesalary 4000.0 properties: - bucket_count -1 column.name.delimiter , columns employeeid,employeename - columns.comments columns.types int:string field.delim | #### A masked pattern was here #### name default.employee_part - numFiles 1 - numRows 0 partition_columns employeesalary partition_columns.types double - rawDataSize 0 - serialization.ddl struct employee_part { i32 employeeid, string employeename} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 105 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns employeeid,employeename @@ -516,10 +496,8 @@ STAGE PLANS: name default.employee_part partition_columns employeesalary partition_columns.types double - serialization.ddl struct employee_part { i32 employeeid, string employeename} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.employee_part name: default.employee_part diff --git a/ql/src/test/results/clientpositive/llap/columnstats_tbllvl.q.out b/ql/src/test/results/clientpositive/llap/columnstats_tbllvl.q.out index 83ae4146f0..f361fda252 100644 --- a/ql/src/test/results/clientpositive/llap/columnstats_tbllvl.q.out +++ b/ql/src/test/results/clientpositive/llap/columnstats_tbllvl.q.out @@ -173,25 +173,17 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite - columns.comments columns.types string:string:string:float:string:string:string:string:int field.delim | #### A masked pattern was here #### name default.uservisits_web_text_none - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7060 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite @@ -200,14 +192,8 @@ STAGE PLANS: field.delim | #### A masked pattern was here #### name default.uservisits_web_text_none - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7060 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.uservisits_web_text_none name: default.uservisits_web_text_none @@ -722,25 +708,17 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite - columns.comments columns.types string:string:string:float:string:string:string:string:int field.delim | #### A masked pattern was here #### name dummydb.uservisits_in_dummy_db - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct uservisits_in_dummy_db { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7060 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite @@ -749,14 +727,8 @@ STAGE PLANS: field.delim | #### A masked pattern was here #### name dummydb.uservisits_in_dummy_db - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct uservisits_in_dummy_db { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7060 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: dummydb.uservisits_in_dummy_db name: dummydb.uservisits_in_dummy_db diff --git a/ql/src/test/results/clientpositive/llap/comments.q.out b/ql/src/test/results/clientpositive/llap/comments.q.out index f04bec9ef8..d0362be681 100644 --- a/ql/src/test/results/clientpositive/llap/comments.q.out +++ b/ql/src/test/results/clientpositive/llap/comments.q.out @@ -151,30 +151,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -182,14 +172,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -231,30 +215,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -262,14 +236,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/llap/constantPropagateForSubQuery.q.out b/ql/src/test/results/clientpositive/llap/constantPropagateForSubQuery.q.out index e20f7beec6..6baff579ee 100644 --- a/ql/src/test/results/clientpositive/llap/constantPropagateForSubQuery.q.out +++ b/ql/src/test/results/clientpositive/llap/constantPropagateForSubQuery.q.out @@ -64,30 +64,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -95,14 +85,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -138,30 +122,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -169,14 +143,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src1 name: default.src1 diff --git a/ql/src/test/results/clientpositive/llap/disable_merge_for_bucketing.q.out b/ql/src/test/results/clientpositive/llap/disable_merge_for_bucketing.q.out index a597f2c075..946f0c4128 100644 --- a/ql/src/test/results/clientpositive/llap/disable_merge_for_bucketing.q.out +++ b/ql/src/test/results/clientpositive/llap/disable_merge_for_bucketing.q.out @@ -66,30 +66,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -97,14 +87,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -130,7 +114,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -140,14 +123,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.bucket2_1_n0 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket2_1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket2_1_n0 TotalFiles: 2 @@ -199,7 +176,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -209,14 +185,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.bucket2_1_n0 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket2_1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket2_1_n0 diff --git a/ql/src/test/results/clientpositive/llap/display_colstats_tbllvl.q.out b/ql/src/test/results/clientpositive/llap/display_colstats_tbllvl.q.out index 92c58fb720..b5fbb21d6e 100644 --- a/ql/src/test/results/clientpositive/llap/display_colstats_tbllvl.q.out +++ b/ql/src/test/results/clientpositive/llap/display_colstats_tbllvl.q.out @@ -198,25 +198,17 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite - columns.comments columns.types string:string:string:float:string:string:string:string:int field.delim | #### A masked pattern was here #### name default.uservisits_web_text_none_n0 - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct uservisits_web_text_none_n0 { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7060 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite @@ -225,14 +217,8 @@ STAGE PLANS: field.delim | #### A masked pattern was here #### name default.uservisits_web_text_none_n0 - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct uservisits_web_text_none_n0 { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7060 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.uservisits_web_text_none_n0 name: default.uservisits_web_text_none_n0 diff --git a/ql/src/test/results/clientpositive/llap/dynamic_partition_skip_default.q.out b/ql/src/test/results/clientpositive/llap/dynamic_partition_skip_default.q.out index a3ec3997d8..aa68fa7ad4 100644 --- a/ql/src/test/results/clientpositive/llap/dynamic_partition_skip_default.q.out +++ b/ql/src/test/results/clientpositive/llap/dynamic_partition_skip_default.q.out @@ -70,30 +70,20 @@ STAGE PLANS: partcol1 1 partcol2 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"intcol":"true"}} - bucket_count -1 column.name.delimiter , columns intcol - columns.comments columns.types string #### A masked pattern was here #### name default.dynamic_part_table - numFiles 1 - numRows 1 partition_columns partcol1/partcol2 partition_columns.types string:string - rawDataSize 1 - serialization.ddl struct dynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns intcol @@ -103,10 +93,8 @@ STAGE PLANS: name default.dynamic_part_table partition_columns partcol1/partcol2 partition_columns.types string:string - serialization.ddl struct dynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dynamic_part_table name: default.dynamic_part_table @@ -148,30 +136,20 @@ STAGE PLANS: partcol1 1 partcol2 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"intcol":"true"}} - bucket_count -1 column.name.delimiter , columns intcol - columns.comments columns.types string #### A masked pattern was here #### name default.dynamic_part_table - numFiles 1 - numRows 1 partition_columns partcol1/partcol2 partition_columns.types string:string - rawDataSize 1 - serialization.ddl struct dynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns intcol @@ -181,10 +159,8 @@ STAGE PLANS: name default.dynamic_part_table partition_columns partcol1/partcol2 partition_columns.types string:string - serialization.ddl struct dynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dynamic_part_table name: default.dynamic_part_table @@ -228,30 +204,20 @@ STAGE PLANS: partcol1 1 partcol2 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"intcol":"true"}} - bucket_count -1 column.name.delimiter , columns intcol - columns.comments columns.types string #### A masked pattern was here #### name default.dynamic_part_table - numFiles 1 - numRows 1 partition_columns partcol1/partcol2 partition_columns.types string:string - rawDataSize 1 - serialization.ddl struct dynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns intcol @@ -261,10 +227,8 @@ STAGE PLANS: name default.dynamic_part_table partition_columns partcol1/partcol2 partition_columns.types string:string - serialization.ddl struct dynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dynamic_part_table name: default.dynamic_part_table @@ -275,30 +239,20 @@ STAGE PLANS: partcol1 1 partcol2 __HIVE_DEFAULT_PARTITION__ properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"intcol":"true"}} - bucket_count -1 column.name.delimiter , columns intcol - columns.comments columns.types string #### A masked pattern was here #### name default.dynamic_part_table - numFiles 1 - numRows 1 partition_columns partcol1/partcol2 partition_columns.types string:string - rawDataSize 1 - serialization.ddl struct dynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns intcol @@ -308,10 +262,8 @@ STAGE PLANS: name default.dynamic_part_table partition_columns partcol1/partcol2 partition_columns.types string:string - serialization.ddl struct dynamic_part_table { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dynamic_part_table name: default.dynamic_part_table diff --git a/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out index e379b14e27..3efb830c4f 100644 --- a/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out +++ b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_reduction.q.out @@ -1784,26 +1784,17 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 #### A masked pattern was here #### name default.srcpart_date_n7 - numFiles 1 - numRows 1000 partition_columns ds partition_columns.types string - rawDataSize 176000 - serialization.ddl struct srcpart_date_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 3054 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1813,10 +1804,8 @@ STAGE PLANS: name default.srcpart_date_n7 partition_columns ds partition_columns.types string - serialization.ddl struct srcpart_date_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.srcpart_date_n7 name: default.srcpart_date_n7 @@ -1828,26 +1817,17 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 #### A masked pattern was here #### name default.srcpart_date_n7 - numFiles 1 - numRows 1000 partition_columns ds partition_columns.types string - rawDataSize 176000 - serialization.ddl struct srcpart_date_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 3054 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1857,10 +1837,8 @@ STAGE PLANS: name default.srcpart_date_n7 partition_columns ds partition_columns.types string - serialization.ddl struct srcpart_date_n7 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.srcpart_date_n7 name: default.srcpart_date_n7 @@ -1925,26 +1903,17 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key1":"true","value1":"true"}} - bucket_count -1 #### A masked pattern was here #### name default.srcpart_small_n3 - numFiles 0 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcpart_small_n3 { string key1, string value1} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,value1 @@ -1954,10 +1923,8 @@ STAGE PLANS: name default.srcpart_small_n3 partition_columns ds partition_columns.types string - serialization.ddl struct srcpart_small_n3 { string key1, string value1} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.srcpart_small_n3 name: default.srcpart_small_n3 @@ -1969,26 +1936,17 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key1":"true","value1":"true"}} - bucket_count -1 #### A masked pattern was here #### name default.srcpart_small_n3 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 3520 - serialization.ddl struct srcpart_small_n3 { string key1, string value1} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 471 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,value1 @@ -1998,10 +1956,8 @@ STAGE PLANS: name default.srcpart_small_n3 partition_columns ds partition_columns.types string - serialization.ddl struct srcpart_small_n3 { string key1, string value1} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.srcpart_small_n3 name: default.srcpart_small_n3 diff --git a/ql/src/test/results/clientpositive/llap/dynamic_semijoin_user_level.q.out b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_user_level.q.out index fec7de2dc1..9bce5436b1 100644 --- a/ql/src/test/results/clientpositive/llap/dynamic_semijoin_user_level.q.out +++ b/ql/src/test/results/clientpositive/llap/dynamic_semijoin_user_level.q.out @@ -984,27 +984,18 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.srcpart_date_n9 - numFiles 1 - numRows 1000 partition_columns ds partition_columns.types string - rawDataSize 176000 - serialization.ddl struct srcpart_date_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 3054 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1015,10 +1006,8 @@ STAGE PLANS: name default.srcpart_date_n9 partition_columns ds partition_columns.types string - serialization.ddl struct srcpart_date_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.srcpart_date_n9 name: default.srcpart_date_n9 @@ -1030,27 +1019,18 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.srcpart_date_n9 - numFiles 1 - numRows 1000 partition_columns ds partition_columns.types string - rawDataSize 176000 - serialization.ddl struct srcpart_date_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 3054 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1061,10 +1041,8 @@ STAGE PLANS: name default.srcpart_date_n9 partition_columns ds partition_columns.types string - serialization.ddl struct srcpart_date_n9 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.srcpart_date_n9 name: default.srcpart_date_n9 @@ -1130,27 +1108,18 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key1":"true","value1":"true"}} - bucket_count -1 #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.srcpart_small_n4 - numFiles 0 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcpart_small_n4 { string key1, string value1} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,value1 @@ -1161,10 +1130,8 @@ STAGE PLANS: name default.srcpart_small_n4 partition_columns ds partition_columns.types string - serialization.ddl struct srcpart_small_n4 { string key1, string value1} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.srcpart_small_n4 name: default.srcpart_small_n4 @@ -1176,27 +1143,18 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key1":"true","value1":"true"}} - bucket_count -1 #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.srcpart_small_n4 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 3520 - serialization.ddl struct srcpart_small_n4 { string key1, string value1} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 471 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,value1 @@ -1207,10 +1165,8 @@ STAGE PLANS: name default.srcpart_small_n4 partition_columns ds partition_columns.types string - serialization.ddl struct srcpart_small_n4 { string key1, string value1} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.srcpart_small_n4 name: default.srcpart_small_n4 diff --git a/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_date.q.out b/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_date.q.out index 2eb4e4ba65..66d3007e64 100644 --- a/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_date.q.out +++ b/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_date.q.out @@ -164,30 +164,20 @@ STAGE PLANS: partition values: d_date_sk 2416945 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"d_date":"true"}} - bucket_count -1 column.name.delimiter , columns d_date - columns.comments columns.types date #### A masked pattern was here #### name default.date_dim_n1 - numFiles 1 - numRows 1 partition_columns d_date_sk partition_columns.types bigint - rawDataSize 56 - serialization.ddl struct date_dim_n1 { date d_date} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 201 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns d_date @@ -197,10 +187,8 @@ STAGE PLANS: name default.date_dim_n1 partition_columns d_date_sk partition_columns.types bigint - serialization.ddl struct date_dim_n1 { date d_date} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.date_dim_n1 name: default.date_dim_n1 @@ -210,30 +198,20 @@ STAGE PLANS: partition values: d_date_sk 2416946 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"d_date":"true"}} - bucket_count -1 column.name.delimiter , columns d_date - columns.comments columns.types date #### A masked pattern was here #### name default.date_dim_n1 - numFiles 1 - numRows 1 partition_columns d_date_sk partition_columns.types bigint - rawDataSize 56 - serialization.ddl struct date_dim_n1 { date d_date} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 201 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns d_date @@ -243,10 +221,8 @@ STAGE PLANS: name default.date_dim_n1 partition_columns d_date_sk partition_columns.types bigint - serialization.ddl struct date_dim_n1 { date d_date} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.date_dim_n1 name: default.date_dim_n1 @@ -256,30 +232,20 @@ STAGE PLANS: partition values: d_date_sk 2416947 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"d_date":"true"}} - bucket_count -1 column.name.delimiter , columns d_date - columns.comments columns.types date #### A masked pattern was here #### name default.date_dim_n1 - numFiles 1 - numRows 1 partition_columns d_date_sk partition_columns.types bigint - rawDataSize 56 - serialization.ddl struct date_dim_n1 { date d_date} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 201 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns d_date @@ -289,10 +255,8 @@ STAGE PLANS: name default.date_dim_n1 partition_columns d_date_sk partition_columns.types bigint - serialization.ddl struct date_dim_n1 { date d_date} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.date_dim_n1 name: default.date_dim_n1 @@ -302,30 +266,20 @@ STAGE PLANS: partition values: d_date_sk 2416948 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"d_date":"true"}} - bucket_count -1 column.name.delimiter , columns d_date - columns.comments columns.types date #### A masked pattern was here #### name default.date_dim_n1 - numFiles 1 - numRows 1 partition_columns d_date_sk partition_columns.types bigint - rawDataSize 56 - serialization.ddl struct date_dim_n1 { date d_date} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 201 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns d_date @@ -335,10 +289,8 @@ STAGE PLANS: name default.date_dim_n1 partition_columns d_date_sk partition_columns.types bigint - serialization.ddl struct date_dim_n1 { date d_date} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.date_dim_n1 name: default.date_dim_n1 diff --git a/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_full.q.out b/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_full.q.out index 3a7d46ade0..cba98dddbe 100644 --- a/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_full.q.out +++ b/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_full.q.out @@ -129,30 +129,20 @@ STAGE PLANS: partition values: year 2000 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d - numFiles 1 - numRows 2 partition_columns year partition_columns.types string - rawDataSize 184 - serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 385 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -162,10 +152,8 @@ STAGE PLANS: name default.loc_orc_1d partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d name: default.loc_orc_1d @@ -175,30 +163,20 @@ STAGE PLANS: partition values: year 2001 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d - numFiles 1 - numRows 4 partition_columns year partition_columns.types string - rawDataSize 368 - serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 406 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -208,10 +186,8 @@ STAGE PLANS: name default.loc_orc_1d partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d name: default.loc_orc_1d @@ -252,30 +228,20 @@ STAGE PLANS: partition values: year 2000 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d - numFiles 1 - numRows 2 partition_columns year partition_columns.types string - rawDataSize 184 - serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 385 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -285,10 +251,8 @@ STAGE PLANS: name default.loc_orc_1d partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d name: default.loc_orc_1d @@ -298,30 +262,20 @@ STAGE PLANS: partition values: year 2001 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d - numFiles 1 - numRows 4 partition_columns year partition_columns.types string - rawDataSize 368 - serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 406 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -331,10 +285,8 @@ STAGE PLANS: name default.loc_orc_1d partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d name: default.loc_orc_1d @@ -469,30 +421,20 @@ STAGE PLANS: year 2000 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 89 - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 293 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -502,10 +444,8 @@ STAGE PLANS: name default.loc_orc_2d partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d name: default.loc_orc_2d @@ -516,30 +456,20 @@ STAGE PLANS: year 2001 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d - numFiles 1 - numRows 3 partition_columns zip/year partition_columns.types int:string - rawDataSize 267 - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 295 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -549,10 +479,8 @@ STAGE PLANS: name default.loc_orc_2d partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d name: default.loc_orc_2d @@ -563,30 +491,20 @@ STAGE PLANS: year 2000 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 88 - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 277 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -596,10 +514,8 @@ STAGE PLANS: name default.loc_orc_2d partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d name: default.loc_orc_2d @@ -610,30 +526,20 @@ STAGE PLANS: year 2001 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 88 - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 277 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -643,10 +549,8 @@ STAGE PLANS: name default.loc_orc_2d partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d name: default.loc_orc_2d @@ -692,30 +596,20 @@ STAGE PLANS: year 2000 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 89 - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 293 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -725,10 +619,8 @@ STAGE PLANS: name default.loc_orc_2d partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d name: default.loc_orc_2d @@ -739,30 +631,20 @@ STAGE PLANS: year 2001 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d - numFiles 1 - numRows 3 partition_columns zip/year partition_columns.types int:string - rawDataSize 267 - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 295 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -772,10 +654,8 @@ STAGE PLANS: name default.loc_orc_2d partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d name: default.loc_orc_2d @@ -786,30 +666,20 @@ STAGE PLANS: year 2000 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 88 - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 277 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -819,10 +689,8 @@ STAGE PLANS: name default.loc_orc_2d partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d name: default.loc_orc_2d @@ -833,30 +701,20 @@ STAGE PLANS: year 2001 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 88 - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 277 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -866,10 +724,8 @@ STAGE PLANS: name default.loc_orc_2d partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d name: default.loc_orc_2d diff --git a/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_partial.q.out b/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_partial.q.out index 3e6024c6a9..de09f9e8ea 100644 --- a/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_partial.q.out +++ b/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_partial.q.out @@ -159,30 +159,20 @@ STAGE PLANS: partition values: year 2000 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 2 partition_columns year partition_columns.types string - rawDataSize 184 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 385 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -192,10 +182,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -205,30 +193,20 @@ STAGE PLANS: partition values: year 2001 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 4 partition_columns year partition_columns.types string - rawDataSize 368 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 407 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -238,10 +216,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -251,30 +227,20 @@ STAGE PLANS: partition values: year 2002 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 6 partition_columns year partition_columns.types string - rawDataSize 570 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 429 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -284,10 +250,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -297,30 +261,20 @@ STAGE PLANS: partition values: year 2003 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 8 partition_columns year partition_columns.types string - rawDataSize 744 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 436 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -330,10 +284,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -378,30 +330,20 @@ STAGE PLANS: partition values: year 2000 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 2 partition_columns year partition_columns.types string - rawDataSize 184 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 385 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -411,10 +353,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -424,30 +364,20 @@ STAGE PLANS: partition values: year 2001 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 4 partition_columns year partition_columns.types string - rawDataSize 368 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 407 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -457,10 +387,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -470,30 +398,20 @@ STAGE PLANS: partition values: year 2002 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 6 partition_columns year partition_columns.types string - rawDataSize 570 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 429 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -503,10 +421,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -516,30 +432,20 @@ STAGE PLANS: partition values: year 2003 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 8 partition_columns year partition_columns.types string - rawDataSize 744 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 436 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -549,10 +455,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -625,30 +529,20 @@ STAGE PLANS: partition values: year 2000 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 2 partition_columns year partition_columns.types string - rawDataSize 184 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 385 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -658,10 +552,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -671,30 +563,20 @@ STAGE PLANS: partition values: year 2001 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 4 partition_columns year partition_columns.types string - rawDataSize 368 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 407 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -704,10 +586,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -717,30 +597,20 @@ STAGE PLANS: partition values: year 2002 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 6 partition_columns year partition_columns.types string - rawDataSize 570 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 429 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -750,10 +620,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -763,30 +631,20 @@ STAGE PLANS: partition values: year 2003 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 8 partition_columns year partition_columns.types string - rawDataSize 744 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 436 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -796,10 +654,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -844,30 +700,20 @@ STAGE PLANS: partition values: year 2000 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 2 partition_columns year partition_columns.types string - rawDataSize 184 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 385 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -877,10 +723,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -890,30 +734,20 @@ STAGE PLANS: partition values: year 2001 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 4 partition_columns year partition_columns.types string - rawDataSize 368 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 407 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -923,10 +757,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -936,30 +768,20 @@ STAGE PLANS: partition values: year 2002 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 6 partition_columns year partition_columns.types string - rawDataSize 570 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 429 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -969,10 +791,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -982,30 +802,20 @@ STAGE PLANS: partition values: year 2003 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,zip - columns.comments columns.types string:int:int #### A masked pattern was here #### name default.loc_orc_1d_n1 - numFiles 1 - numRows 8 partition_columns year partition_columns.types string - rawDataSize 744 - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 436 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,zip @@ -1015,10 +825,8 @@ STAGE PLANS: name default.loc_orc_1d_n1 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n1 { string state, i32 locid, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n1 name: default.loc_orc_1d_n1 @@ -1160,30 +968,20 @@ STAGE PLANS: year 2001 zip 43201 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 90 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 298 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1193,10 +991,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1207,30 +1003,20 @@ STAGE PLANS: year 2002 zip 43201 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 2 partition_columns zip/year partition_columns.types int:string - rawDataSize 182 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 307 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1240,10 +1026,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1254,30 +1038,20 @@ STAGE PLANS: year 2003 zip 43201 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 3 partition_columns zip/year partition_columns.types int:string - rawDataSize 267 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 314 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1287,10 +1061,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1301,30 +1073,20 @@ STAGE PLANS: year 2000 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 89 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 293 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1334,10 +1096,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1348,30 +1108,20 @@ STAGE PLANS: year 2001 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 2 partition_columns zip/year partition_columns.types int:string - rawDataSize 176 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 288 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1381,10 +1131,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1395,30 +1143,20 @@ STAGE PLANS: year 2002 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 91 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 300 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1428,10 +1166,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1442,30 +1178,20 @@ STAGE PLANS: year 2003 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 2 partition_columns zip/year partition_columns.types int:string - rawDataSize 180 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 309 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1475,10 +1201,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1489,30 +1213,20 @@ STAGE PLANS: year 2000 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 88 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 277 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1522,10 +1236,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1536,30 +1248,20 @@ STAGE PLANS: year 2001 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 88 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 277 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1569,10 +1271,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1583,30 +1283,20 @@ STAGE PLANS: year 2002 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 3 partition_columns zip/year partition_columns.types int:string - rawDataSize 273 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 310 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1616,10 +1306,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1630,30 +1318,20 @@ STAGE PLANS: year 2003 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 3 partition_columns zip/year partition_columns.types int:string - rawDataSize 264 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 307 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1663,10 +1341,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1726,30 +1402,20 @@ STAGE PLANS: year 2001 zip 43201 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 90 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 298 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1759,10 +1425,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1773,30 +1437,20 @@ STAGE PLANS: year 2002 zip 43201 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 2 partition_columns zip/year partition_columns.types int:string - rawDataSize 182 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 307 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1806,10 +1460,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1820,30 +1472,20 @@ STAGE PLANS: year 2003 zip 43201 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 3 partition_columns zip/year partition_columns.types int:string - rawDataSize 267 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 314 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1853,10 +1495,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1867,30 +1507,20 @@ STAGE PLANS: year 2000 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 89 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 293 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1900,10 +1530,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1914,30 +1542,20 @@ STAGE PLANS: year 2001 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 2 partition_columns zip/year partition_columns.types int:string - rawDataSize 176 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 288 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1947,10 +1565,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -1961,30 +1577,20 @@ STAGE PLANS: year 2002 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 91 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 300 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -1994,10 +1600,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -2008,30 +1612,20 @@ STAGE PLANS: year 2003 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 2 partition_columns zip/year partition_columns.types int:string - rawDataSize 180 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 309 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -2041,10 +1635,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -2055,30 +1647,20 @@ STAGE PLANS: year 2000 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 88 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 277 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -2088,10 +1670,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -2102,30 +1682,20 @@ STAGE PLANS: year 2001 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 88 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 277 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -2135,10 +1705,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -2149,30 +1717,20 @@ STAGE PLANS: year 2002 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 3 partition_columns zip/year partition_columns.types int:string - rawDataSize 273 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 310 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -2182,10 +1740,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 @@ -2196,30 +1752,20 @@ STAGE PLANS: year 2003 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid - columns.comments columns.types string:int #### A masked pattern was here #### name default.loc_orc_2d_n1 - numFiles 1 - numRows 3 partition_columns zip/year partition_columns.types int:string - rawDataSize 264 - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 307 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid @@ -2229,10 +1775,8 @@ STAGE PLANS: name default.loc_orc_2d_n1 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n1 { string state, i32 locid} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n1 name: default.loc_orc_2d_n1 diff --git a/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_partial_ndv.q.out b/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_partial_ndv.q.out index e522926a90..383b2dbc12 100644 --- a/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_partial_ndv.q.out +++ b/ql/src/test/results/clientpositive/llap/extrapolate_part_stats_partial_ndv.q.out @@ -283,30 +283,20 @@ STAGE PLANS: partition values: year 2000 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt,zip - columns.comments columns.types string:double:decimal(10,0):int #### A masked pattern was here #### name default.loc_orc_1d_n0 - numFiles 1 - numRows 2 partition_columns year partition_columns.types string - rawDataSize 416 - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 544 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt,zip @@ -316,10 +306,8 @@ STAGE PLANS: name default.loc_orc_1d_n0 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n0 name: default.loc_orc_1d_n0 @@ -329,30 +317,20 @@ STAGE PLANS: partition values: year 2001 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt,zip - columns.comments columns.types string:double:decimal(10,0):int #### A masked pattern was here #### name default.loc_orc_1d_n0 - numFiles 1 - numRows 4 partition_columns year partition_columns.types string - rawDataSize 832 - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 583 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt,zip @@ -362,10 +340,8 @@ STAGE PLANS: name default.loc_orc_1d_n0 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n0 name: default.loc_orc_1d_n0 @@ -375,30 +351,20 @@ STAGE PLANS: partition values: year 2002 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt,zip - columns.comments columns.types string:double:decimal(10,0):int #### A masked pattern was here #### name default.loc_orc_1d_n0 - numFiles 1 - numRows 6 partition_columns year partition_columns.types string - rawDataSize 1266 - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 599 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt,zip @@ -408,10 +374,8 @@ STAGE PLANS: name default.loc_orc_1d_n0 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n0 name: default.loc_orc_1d_n0 @@ -421,30 +385,20 @@ STAGE PLANS: partition values: year 2003 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt,zip - columns.comments columns.types string:double:decimal(10,0):int #### A masked pattern was here #### name default.loc_orc_1d_n0 - numFiles 1 - numRows 8 partition_columns year partition_columns.types string - rawDataSize 1672 - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 622 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt,zip @@ -454,10 +408,8 @@ STAGE PLANS: name default.loc_orc_1d_n0 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n0 name: default.loc_orc_1d_n0 @@ -674,30 +626,20 @@ STAGE PLANS: partition values: year 2000 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt,zip - columns.comments columns.types string:double:decimal(10,0):int #### A masked pattern was here #### name default.loc_orc_1d_n0 - numFiles 1 - numRows 2 partition_columns year partition_columns.types string - rawDataSize 416 - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 544 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt,zip @@ -707,10 +649,8 @@ STAGE PLANS: name default.loc_orc_1d_n0 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n0 name: default.loc_orc_1d_n0 @@ -720,30 +660,20 @@ STAGE PLANS: partition values: year 2001 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt,zip - columns.comments columns.types string:double:decimal(10,0):int #### A masked pattern was here #### name default.loc_orc_1d_n0 - numFiles 1 - numRows 4 partition_columns year partition_columns.types string - rawDataSize 832 - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 583 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt,zip @@ -753,10 +683,8 @@ STAGE PLANS: name default.loc_orc_1d_n0 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n0 name: default.loc_orc_1d_n0 @@ -766,30 +694,20 @@ STAGE PLANS: partition values: year 2002 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt,zip - columns.comments columns.types string:double:decimal(10,0):int #### A masked pattern was here #### name default.loc_orc_1d_n0 - numFiles 1 - numRows 6 partition_columns year partition_columns.types string - rawDataSize 1266 - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 599 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt,zip @@ -799,10 +717,8 @@ STAGE PLANS: name default.loc_orc_1d_n0 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n0 name: default.loc_orc_1d_n0 @@ -812,30 +728,20 @@ STAGE PLANS: partition values: year 2003 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true","zip":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt,zip - columns.comments columns.types string:double:decimal(10,0):int #### A masked pattern was here #### name default.loc_orc_1d_n0 - numFiles 1 - numRows 8 partition_columns year partition_columns.types string - rawDataSize 1672 - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 622 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt,zip @@ -845,10 +751,8 @@ STAGE PLANS: name default.loc_orc_1d_n0 partition_columns year partition_columns.types string - serialization.ddl struct loc_orc_1d_n0 { string state, double locid, decimal(10,0) cnt, i32 zip} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_1d_n0 name: default.loc_orc_1d_n0 @@ -1115,30 +1019,20 @@ STAGE PLANS: year 2001 zip 43201 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt - columns.comments columns.types string:int:decimal(10,0) #### A masked pattern was here #### name default.loc_orc_2d_n0 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 202 - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 401 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt @@ -1148,10 +1042,8 @@ STAGE PLANS: name default.loc_orc_2d_n0 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n0 name: default.loc_orc_2d_n0 @@ -1162,30 +1054,20 @@ STAGE PLANS: year 2002 zip 43201 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt - columns.comments columns.types string:int:decimal(10,0) #### A masked pattern was here #### name default.loc_orc_2d_n0 - numFiles 1 - numRows 2 partition_columns zip/year partition_columns.types int:string - rawDataSize 406 - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 425 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt @@ -1195,10 +1077,8 @@ STAGE PLANS: name default.loc_orc_2d_n0 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n0 name: default.loc_orc_2d_n0 @@ -1209,30 +1089,20 @@ STAGE PLANS: year 2003 zip 43201 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt - columns.comments columns.types string:int:decimal(10,0) #### A masked pattern was here #### name default.loc_orc_2d_n0 - numFiles 1 - numRows 3 partition_columns zip/year partition_columns.types int:string - rawDataSize 603 - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 439 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt @@ -1242,10 +1112,8 @@ STAGE PLANS: name default.loc_orc_2d_n0 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n0 name: default.loc_orc_2d_n0 @@ -1256,30 +1124,20 @@ STAGE PLANS: year 2000 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt - columns.comments columns.types string:int:decimal(10,0) #### A masked pattern was here #### name default.loc_orc_2d_n0 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 201 - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 401 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt @@ -1289,10 +1147,8 @@ STAGE PLANS: name default.loc_orc_2d_n0 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n0 name: default.loc_orc_2d_n0 @@ -1303,30 +1159,20 @@ STAGE PLANS: year 2001 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt - columns.comments columns.types string:int:decimal(10,0) #### A masked pattern was here #### name default.loc_orc_2d_n0 - numFiles 1 - numRows 2 partition_columns zip/year partition_columns.types int:string - rawDataSize 400 - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 409 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt @@ -1336,10 +1182,8 @@ STAGE PLANS: name default.loc_orc_2d_n0 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n0 name: default.loc_orc_2d_n0 @@ -1350,30 +1194,20 @@ STAGE PLANS: year 2002 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt - columns.comments columns.types string:int:decimal(10,0) #### A masked pattern was here #### name default.loc_orc_2d_n0 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 203 - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 404 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt @@ -1383,10 +1217,8 @@ STAGE PLANS: name default.loc_orc_2d_n0 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n0 name: default.loc_orc_2d_n0 @@ -1397,30 +1229,20 @@ STAGE PLANS: year 2003 zip 94086 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt - columns.comments columns.types string:int:decimal(10,0) #### A masked pattern was here #### name default.loc_orc_2d_n0 - numFiles 1 - numRows 2 partition_columns zip/year partition_columns.types int:string - rawDataSize 404 - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 425 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt @@ -1430,10 +1252,8 @@ STAGE PLANS: name default.loc_orc_2d_n0 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n0 name: default.loc_orc_2d_n0 @@ -1444,30 +1264,20 @@ STAGE PLANS: year 2000 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt - columns.comments columns.types string:int:decimal(10,0) #### A masked pattern was here #### name default.loc_orc_2d_n0 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 200 - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 380 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt @@ -1477,10 +1287,8 @@ STAGE PLANS: name default.loc_orc_2d_n0 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n0 name: default.loc_orc_2d_n0 @@ -1491,30 +1299,20 @@ STAGE PLANS: year 2001 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt - columns.comments columns.types string:int:decimal(10,0) #### A masked pattern was here #### name default.loc_orc_2d_n0 - numFiles 1 - numRows 1 partition_columns zip/year partition_columns.types int:string - rawDataSize 200 - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 375 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt @@ -1524,10 +1322,8 @@ STAGE PLANS: name default.loc_orc_2d_n0 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n0 name: default.loc_orc_2d_n0 @@ -1538,30 +1334,20 @@ STAGE PLANS: year 2002 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt - columns.comments columns.types string:int:decimal(10,0) #### A masked pattern was here #### name default.loc_orc_2d_n0 - numFiles 1 - numRows 3 partition_columns zip/year partition_columns.types int:string - rawDataSize 609 - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 429 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt @@ -1571,10 +1357,8 @@ STAGE PLANS: name default.loc_orc_2d_n0 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n0 name: default.loc_orc_2d_n0 @@ -1585,30 +1369,20 @@ STAGE PLANS: year 2003 zip 94087 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","locid":"true","state":"true"}} - bucket_count -1 column.name.delimiter , columns state,locid,cnt - columns.comments columns.types string:int:decimal(10,0) #### A masked pattern was here #### name default.loc_orc_2d_n0 - numFiles 1 - numRows 3 partition_columns zip/year partition_columns.types int:string - rawDataSize 600 - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 430 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns state,locid,cnt @@ -1618,10 +1392,8 @@ STAGE PLANS: name default.loc_orc_2d_n0 partition_columns zip/year partition_columns.types int:string - serialization.ddl struct loc_orc_2d_n0 { string state, i32 locid, decimal(10,0) cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.loc_orc_2d_n0 name: default.loc_orc_2d_n0 diff --git a/ql/src/test/results/clientpositive/llap/filter_aggr.q.out b/ql/src/test/results/clientpositive/llap/filter_aggr.q.out index 49472ccd6e..11c946bb7f 100644 --- a/ql/src/test/results/clientpositive/llap/filter_aggr.q.out +++ b/ql/src/test/results/clientpositive/llap/filter_aggr.q.out @@ -74,30 +74,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -105,14 +95,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/llap/filter_join_breaktask.q.out b/ql/src/test/results/clientpositive/llap/filter_join_breaktask.q.out index a143c0b6f7..4956d54638 100644 --- a/ql/src/test/results/clientpositive/llap/filter_join_breaktask.q.out +++ b/ql/src/test/results/clientpositive/llap/filter_join_breaktask.q.out @@ -95,30 +95,20 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.filter_join_breaktask - numFiles 1 - numRows 25 partition_columns ds partition_columns.types string - rawDataSize 211 - serialization.ddl struct filter_join_breaktask { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 236 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -128,10 +118,8 @@ STAGE PLANS: name default.filter_join_breaktask partition_columns ds partition_columns.types string - serialization.ddl struct filter_join_breaktask { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.filter_join_breaktask name: default.filter_join_breaktask @@ -176,30 +164,20 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.filter_join_breaktask - numFiles 1 - numRows 25 partition_columns ds partition_columns.types string - rawDataSize 211 - serialization.ddl struct filter_join_breaktask { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 236 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -209,10 +187,8 @@ STAGE PLANS: name default.filter_join_breaktask partition_columns ds partition_columns.types string - serialization.ddl struct filter_join_breaktask { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.filter_join_breaktask name: default.filter_join_breaktask @@ -256,30 +232,20 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.filter_join_breaktask - numFiles 1 - numRows 25 partition_columns ds partition_columns.types string - rawDataSize 211 - serialization.ddl struct filter_join_breaktask { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 236 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -289,10 +255,8 @@ STAGE PLANS: name default.filter_join_breaktask partition_columns ds partition_columns.types string - serialization.ddl struct filter_join_breaktask { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.filter_join_breaktask name: default.filter_join_breaktask diff --git a/ql/src/test/results/clientpositive/llap/filter_union.q.out b/ql/src/test/results/clientpositive/llap/filter_union.q.out index 50761ebcf6..2e472ef840 100644 --- a/ql/src/test/results/clientpositive/llap/filter_union.q.out +++ b/ql/src/test/results/clientpositive/llap/filter_union.q.out @@ -87,30 +87,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -118,14 +108,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -170,30 +154,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -201,14 +175,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/llap/groupby_map_ppr.q.out b/ql/src/test/results/clientpositive/llap/groupby_map_ppr.q.out index 09b9e2f37b..1f8d703728 100644 --- a/ql/src/test/results/clientpositive/llap/groupby_map_ppr.q.out +++ b/ql/src/test/results/clientpositive/llap/groupby_map_ppr.q.out @@ -90,30 +90,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -123,10 +113,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -139,30 +127,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -172,10 +150,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -208,8 +184,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,c1,c2 @@ -217,14 +191,8 @@ STAGE PLANS: columns.types string:int:string #### A masked pattern was here #### name default.dest1_n144 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n144 { string key, i32 c1, string c2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n144 TotalFiles: 1 @@ -295,8 +263,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,c1,c2 @@ -304,14 +270,8 @@ STAGE PLANS: columns.types string:int:string #### A masked pattern was here #### name default.dest1_n144 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n144 { string key, i32 c1, string c2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n144 diff --git a/ql/src/test/results/clientpositive/llap/groupby_map_ppr_multi_distinct.q.out b/ql/src/test/results/clientpositive/llap/groupby_map_ppr_multi_distinct.q.out index c5830c769d..8f93eb0dcf 100644 --- a/ql/src/test/results/clientpositive/llap/groupby_map_ppr_multi_distinct.q.out +++ b/ql/src/test/results/clientpositive/llap/groupby_map_ppr_multi_distinct.q.out @@ -90,30 +90,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -123,10 +113,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -139,30 +127,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -172,10 +150,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -208,8 +184,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,c1,c2,c3,c4 @@ -217,14 +191,8 @@ STAGE PLANS: columns.types string:int:string:int:int #### A masked pattern was here #### name default.dest1_n174 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n174 { string key, i32 c1, string c2, i32 c3, i32 c4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n174 TotalFiles: 1 @@ -295,8 +263,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,c1,c2,c3,c4 @@ -304,14 +270,8 @@ STAGE PLANS: columns.types string:int:string:int:int #### A masked pattern was here #### name default.dest1_n174 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n174 { string key, i32 c1, string c2, i32 c3, i32 c4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n174 diff --git a/ql/src/test/results/clientpositive/llap/groupby_ppr.q.out b/ql/src/test/results/clientpositive/llap/groupby_ppr.q.out index b54f27c8bc..9b857193a2 100644 --- a/ql/src/test/results/clientpositive/llap/groupby_ppr.q.out +++ b/ql/src/test/results/clientpositive/llap/groupby_ppr.q.out @@ -82,30 +82,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -115,10 +105,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -131,30 +119,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -164,10 +142,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -200,8 +176,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,c1,c2 @@ -209,14 +183,8 @@ STAGE PLANS: columns.types string:int:string #### A masked pattern was here #### name default.dest1_n79 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n79 { string key, i32 c1, string c2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n79 TotalFiles: 1 @@ -281,8 +249,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,c1,c2 @@ -290,14 +256,8 @@ STAGE PLANS: columns.types string:int:string #### A masked pattern was here #### name default.dest1_n79 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n79 { string key, i32 c1, string c2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n79 diff --git a/ql/src/test/results/clientpositive/llap/groupby_ppr_multi_distinct.q.out b/ql/src/test/results/clientpositive/llap/groupby_ppr_multi_distinct.q.out index c307d9c25a..0f12ff8348 100644 --- a/ql/src/test/results/clientpositive/llap/groupby_ppr_multi_distinct.q.out +++ b/ql/src/test/results/clientpositive/llap/groupby_ppr_multi_distinct.q.out @@ -82,30 +82,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -115,10 +105,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -131,30 +119,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -164,10 +142,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -200,8 +176,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,c1,c2,c3,c4 @@ -209,14 +183,8 @@ STAGE PLANS: columns.types string:int:string:int:int #### A masked pattern was here #### name default.dest1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1 { string key, i32 c1, string c2, i32 c3, i32 c4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 TotalFiles: 1 @@ -281,8 +249,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,c1,c2,c3,c4 @@ -290,14 +256,8 @@ STAGE PLANS: columns.types string:int:string:int:int #### A masked pattern was here #### name default.dest1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1 { string key, i32 c1, string c2, i32 c3, i32 c4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 @@ -426,30 +386,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -459,10 +409,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -475,30 +423,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -508,10 +446,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -544,8 +480,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,c1,c2,c3,c4 @@ -553,14 +487,8 @@ STAGE PLANS: columns.types string:int:string:int:int #### A masked pattern was here #### name default.dest1 - numFiles 1 - numRows 10 - rawDataSize 184 - serialization.ddl struct dest1 { string key, i32 c1, string c2, i32 c3, i32 c4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 194 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 TotalFiles: 1 @@ -629,8 +557,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"c1":"true","c2":"true","c3":"true","c4":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,c1,c2,c3,c4 @@ -638,14 +564,8 @@ STAGE PLANS: columns.types string:int:string:int:int #### A masked pattern was here #### name default.dest1 - numFiles 1 - numRows 10 - rawDataSize 184 - serialization.ddl struct dest1 { string key, i32 c1, string c2, i32 c3, i32 c4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 194 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1 diff --git a/ql/src/test/results/clientpositive/llap/groupby_sort_1_23.q.out b/ql/src/test/results/clientpositive/llap/groupby_sort_1_23.q.out index 023929da92..4228454958 100644 --- a/ql/src/test/results/clientpositive/llap/groupby_sort_1_23.q.out +++ b/ql/src/test/results/clientpositive/llap/groupby_sort_1_23.q.out @@ -95,8 +95,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -104,14 +102,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 TotalFiles: 1 @@ -147,31 +139,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -182,14 +165,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -241,8 +218,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -250,14 +225,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 @@ -373,31 +342,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -408,14 +368,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -447,8 +401,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,cnt @@ -456,14 +408,8 @@ STAGE PLANS: columns.types int:string:int #### A masked pattern was here #### name default.outputtbl2_n5 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl2_n5 { i32 key1, string key2, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl2_n5 TotalFiles: 1 @@ -534,8 +480,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,cnt @@ -543,14 +487,8 @@ STAGE PLANS: columns.types int:string:int #### A masked pattern was here #### name default.outputtbl2_n5 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl2_n5 { i32 key1, string key2, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl2_n5 @@ -652,8 +590,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -661,14 +597,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 TotalFiles: 1 @@ -704,31 +634,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -739,14 +660,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -798,8 +713,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -807,14 +720,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 @@ -914,8 +821,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -923,14 +828,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 TotalFiles: 1 @@ -966,31 +865,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -1001,14 +891,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -1060,8 +944,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -1069,14 +951,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 @@ -1184,8 +1060,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,cnt @@ -1193,14 +1067,8 @@ STAGE PLANS: columns.types int:int:int #### A masked pattern was here #### name default.outputtbl3_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl3_n2 TotalFiles: 1 @@ -1236,31 +1104,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -1271,14 +1130,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -1330,8 +1183,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,cnt @@ -1339,14 +1190,8 @@ STAGE PLANS: columns.types int:int:int #### A masked pattern was here #### name default.outputtbl3_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl3_n2 @@ -1463,31 +1308,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -1498,14 +1334,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -1537,8 +1367,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -1546,14 +1374,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n2 TotalFiles: 1 @@ -1624,8 +1446,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -1633,14 +1453,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n2 @@ -1751,31 +1565,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -1786,14 +1591,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -1825,8 +1624,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,cnt @@ -1834,14 +1631,8 @@ STAGE PLANS: columns.types int:int:int #### A masked pattern was here #### name default.outputtbl3_n2 - numFiles 1 - numRows 5 - rawDataSize 25 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl3_n2 TotalFiles: 1 @@ -1912,8 +1703,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,cnt @@ -1921,14 +1710,8 @@ STAGE PLANS: columns.types int:int:int #### A masked pattern was here #### name default.outputtbl3_n2 - numFiles 1 - numRows 5 - rawDataSize 25 - serialization.ddl struct outputtbl3_n2 { i32 key1, i32 key2, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl3_n2 @@ -2051,31 +1834,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -2086,14 +1860,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -2125,8 +1893,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -2134,14 +1900,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 TotalFiles: 1 @@ -2212,8 +1972,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -2221,14 +1979,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 @@ -2346,8 +2098,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -2355,14 +2105,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 2 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 TotalFiles: 1 @@ -2398,31 +2142,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -2433,14 +2168,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -2478,8 +2207,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -2487,14 +2214,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 2 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 TotalFiles: 1 @@ -2530,31 +2251,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -2565,14 +2277,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -2626,8 +2332,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -2635,14 +2339,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 2 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 @@ -2769,8 +2467,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -2778,14 +2474,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 2 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 TotalFiles: 1 @@ -2821,31 +2511,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -2856,14 +2537,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -2908,31 +2583,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -2943,14 +2609,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -3019,8 +2679,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -3028,14 +2686,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 2 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 TotalFiles: 1 @@ -3075,8 +2727,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -3084,14 +2734,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 2 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 @@ -3212,31 +2856,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -3247,14 +2882,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -3301,8 +2930,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -3310,14 +2937,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 3 - numRows 10 - rawDataSize 32 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 42 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 TotalFiles: 1 @@ -3351,31 +2972,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -3386,14 +2998,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -3445,8 +3051,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -3454,14 +3058,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 3 - numRows 10 - rawDataSize 32 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 42 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 @@ -3586,31 +3184,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -3621,14 +3210,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -3674,31 +3257,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -3709,14 +3283,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n80 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n80 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n80 name: default.t1_n80 @@ -3877,31 +3445,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val @@ -3912,14 +3471,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t2_n49 name: default.t2_n49 @@ -3951,8 +3504,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -3960,14 +3511,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 1 - numRows 4 - rawDataSize 12 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 16 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 TotalFiles: 1 @@ -4038,8 +3583,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -4047,14 +3590,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n18 - numFiles 1 - numRows 4 - rawDataSize 12 - serialization.ddl struct outputtbl1_n18 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 16 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n18 @@ -4154,8 +3691,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -4163,14 +3698,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n2 - numFiles 2 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n2 TotalFiles: 1 @@ -4206,31 +3735,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val @@ -4241,14 +3761,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t2_n49 name: default.t2_n49 @@ -4300,8 +3814,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -4309,14 +3821,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n2 - numFiles 2 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n2 @@ -4427,8 +3933,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,key4,cnt @@ -4436,14 +3940,8 @@ STAGE PLANS: columns.types int:int:string:int:int #### A masked pattern was here #### name default.outputtbl5_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n2 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl5_n2 TotalFiles: 1 @@ -4479,31 +3977,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val @@ -4514,14 +4003,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t2_n49 name: default.t2_n49 @@ -4573,8 +4056,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,key4,cnt @@ -4582,14 +4063,8 @@ STAGE PLANS: columns.types int:int:string:int:int #### A masked pattern was here #### name default.outputtbl5_n2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n2 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl5_n2 @@ -4699,8 +4174,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -4708,14 +4181,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n2 - numFiles 2 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n2 TotalFiles: 1 @@ -4751,31 +4218,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val @@ -4786,14 +4244,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t2_n49 name: default.t2_n49 @@ -4845,8 +4297,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -4854,14 +4304,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n2 - numFiles 2 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n2 @@ -4978,8 +4422,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -4987,14 +4429,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n2 - numFiles 2 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n2 TotalFiles: 1 @@ -5030,31 +4466,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val @@ -5065,14 +4492,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t2_n49 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n49 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t2_n49 name: default.t2_n49 @@ -5124,8 +4545,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -5133,14 +4552,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n2 - numFiles 2 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n2 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n2 diff --git a/ql/src/test/results/clientpositive/llap/groupby_sort_6.q.out b/ql/src/test/results/clientpositive/llap/groupby_sort_6.q.out index 68f196cd03..08d94dc5c5 100644 --- a/ql/src/test/results/clientpositive/llap/groupby_sort_6.q.out +++ b/ql/src/test/results/clientpositive/llap/groupby_sort_6.q.out @@ -106,8 +106,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -115,14 +113,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n15 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n15 TotalFiles: 1 @@ -193,8 +185,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -202,14 +192,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n15 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n15 @@ -344,8 +328,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -353,14 +335,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n15 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n15 TotalFiles: 1 @@ -431,8 +407,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -440,14 +414,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n15 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n15 @@ -556,29 +524,20 @@ STAGE PLANS: partition values: ds 2 properties: - bucket_count -1 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n61 - numFiles 1 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct t1_n61 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,val @@ -588,10 +547,8 @@ STAGE PLANS: name default.t1_n61 partition_columns ds partition_columns.types string - serialization.ddl struct t1_n61 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n61 name: default.t1_n61 @@ -623,8 +580,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -632,14 +587,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n15 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n15 TotalFiles: 1 @@ -710,8 +659,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -719,14 +666,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n15 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n15 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n15 diff --git a/ql/src/test/results/clientpositive/llap/groupby_sort_skew_1_23.q.out b/ql/src/test/results/clientpositive/llap/groupby_sort_skew_1_23.q.out index 5b9ee467a0..0e528ffa37 100644 --- a/ql/src/test/results/clientpositive/llap/groupby_sort_skew_1_23.q.out +++ b/ql/src/test/results/clientpositive/llap/groupby_sort_skew_1_23.q.out @@ -95,8 +95,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -104,14 +102,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 TotalFiles: 1 @@ -147,31 +139,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -182,14 +165,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -241,8 +218,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -250,14 +225,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 @@ -374,31 +343,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -409,14 +369,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -469,8 +423,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,cnt @@ -478,14 +430,8 @@ STAGE PLANS: columns.types int:string:int #### A masked pattern was here #### name default.outputtbl2_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl2_n3 { i32 key1, string key2, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl2_n3 TotalFiles: 1 @@ -556,8 +502,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,cnt @@ -565,14 +509,8 @@ STAGE PLANS: columns.types int:string:int #### A masked pattern was here #### name default.outputtbl2_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl2_n3 { i32 key1, string key2, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl2_n3 @@ -674,8 +612,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -683,14 +619,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 TotalFiles: 1 @@ -726,31 +656,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -761,14 +682,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -820,8 +735,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -829,14 +742,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 @@ -936,8 +843,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -945,14 +850,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 TotalFiles: 1 @@ -988,31 +887,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -1023,14 +913,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -1082,8 +966,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -1091,14 +973,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 @@ -1206,8 +1082,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,cnt @@ -1215,14 +1089,8 @@ STAGE PLANS: columns.types int:int:int #### A masked pattern was here #### name default.outputtbl3_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl3_n1 TotalFiles: 1 @@ -1258,31 +1126,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -1293,14 +1152,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -1352,8 +1205,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,cnt @@ -1361,14 +1212,8 @@ STAGE PLANS: columns.types int:int:int #### A masked pattern was here #### name default.outputtbl3_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl3_n1 @@ -1486,31 +1331,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -1521,14 +1357,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -1581,8 +1411,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -1590,14 +1418,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n1 TotalFiles: 1 @@ -1668,8 +1490,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -1677,14 +1497,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n1 @@ -1796,31 +1610,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -1831,14 +1636,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -1891,8 +1690,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,cnt @@ -1900,14 +1697,8 @@ STAGE PLANS: columns.types int:int:int #### A masked pattern was here #### name default.outputtbl3_n1 - numFiles 1 - numRows 5 - rawDataSize 25 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl3_n1 TotalFiles: 1 @@ -1978,8 +1769,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,cnt @@ -1987,14 +1776,8 @@ STAGE PLANS: columns.types int:int:int #### A masked pattern was here #### name default.outputtbl3_n1 - numFiles 1 - numRows 5 - rawDataSize 25 - serialization.ddl struct outputtbl3_n1 { i32 key1, i32 key2, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl3_n1 @@ -2118,31 +1901,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -2153,14 +1927,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -2213,8 +1981,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -2222,14 +1988,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 TotalFiles: 1 @@ -2300,8 +2060,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -2309,14 +2067,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 1 - numRows 5 - rawDataSize 15 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 20 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 @@ -2434,8 +2186,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -2443,14 +2193,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 2 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 TotalFiles: 1 @@ -2486,31 +2230,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -2521,14 +2256,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -2566,8 +2295,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -2575,14 +2302,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 2 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 TotalFiles: 1 @@ -2618,31 +2339,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -2653,14 +2365,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -2714,8 +2420,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -2723,14 +2427,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 2 - numRows 5 - rawDataSize 17 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 22 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 @@ -2858,8 +2556,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -2867,14 +2563,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 2 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 TotalFiles: 1 @@ -2910,31 +2600,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -2945,14 +2626,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -2997,31 +2672,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -3032,14 +2698,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -3129,8 +2789,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -3138,14 +2796,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 2 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 TotalFiles: 1 @@ -3185,8 +2837,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -3194,14 +2844,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 2 - numRows 10 - rawDataSize 30 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 40 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 @@ -3322,31 +2966,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -3357,14 +2992,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -3411,8 +3040,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -3420,14 +3047,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 3 - numRows 10 - rawDataSize 32 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 42 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 TotalFiles: 1 @@ -3461,31 +3082,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -3496,14 +3108,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -3555,8 +3161,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -3564,14 +3168,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 3 - numRows 10 - rawDataSize 32 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 42 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 @@ -3697,31 +3295,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -3732,14 +3321,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -3785,31 +3368,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key @@ -3820,14 +3394,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t1_n56 - numFiles 1 - numRows 6 - rawDataSize 24 - serialization.ddl struct t1_n56 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_n56 name: default.t1_n56 @@ -4010,31 +3578,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val @@ -4045,14 +3604,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t2_n34 name: default.t2_n34 @@ -4105,8 +3658,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -4114,14 +3665,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 1 - numRows 4 - rawDataSize 12 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 16 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 TotalFiles: 1 @@ -4192,8 +3737,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,cnt @@ -4201,14 +3744,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.outputtbl1_n13 - numFiles 1 - numRows 4 - rawDataSize 12 - serialization.ddl struct outputtbl1_n13 { i32 key, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 16 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl1_n13 @@ -4308,8 +3845,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -4317,14 +3852,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n1 - numFiles 2 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n1 TotalFiles: 1 @@ -4360,31 +3889,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val @@ -4395,14 +3915,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t2_n34 name: default.t2_n34 @@ -4454,8 +3968,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -4463,14 +3975,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n1 - numFiles 2 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n1 @@ -4581,8 +4087,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,key4,cnt @@ -4590,14 +4094,8 @@ STAGE PLANS: columns.types int:int:string:int:int #### A masked pattern was here #### name default.outputtbl5_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n1 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl5_n1 TotalFiles: 1 @@ -4633,31 +4131,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val @@ -4668,14 +4157,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t2_n34 name: default.t2_n34 @@ -4727,8 +4210,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true","key4":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,key4,cnt @@ -4736,14 +4217,8 @@ STAGE PLANS: columns.types int:int:string:int:int #### A masked pattern was here #### name default.outputtbl5_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct outputtbl5_n1 { i32 key1, i32 key2, string key3, i32 key4, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl5_n1 @@ -4853,8 +4328,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -4862,14 +4335,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n1 - numFiles 2 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n1 TotalFiles: 1 @@ -4905,31 +4372,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val @@ -4940,14 +4398,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t2_n34 name: default.t2_n34 @@ -4999,8 +4451,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -5008,14 +4458,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n1 - numFiles 2 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n1 @@ -5132,8 +4576,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -5141,14 +4583,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n1 - numFiles 2 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n1 TotalFiles: 1 @@ -5184,31 +4620,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val bucketing_version 2 column.name.delimiter , columns key,val - columns.comments columns.types string:string #### A masked pattern was here #### name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 2 bucket_field_name key,val @@ -5219,14 +4646,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.t2_n34 - numFiles 2 - numRows 6 - rawDataSize 24 - serialization.ddl struct t2_n34 { string key, string val} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 30 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t2_n34 name: default.t2_n34 @@ -5278,8 +4699,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cnt":"true","key1":"true","key2":"true","key3":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,key2,key3,cnt @@ -5287,14 +4706,8 @@ STAGE PLANS: columns.types int:int:string:int #### A masked pattern was here #### name default.outputtbl4_n1 - numFiles 2 - numRows 6 - rawDataSize 48 - serialization.ddl struct outputtbl4_n1 { i32 key1, i32 key2, string key3, i32 cnt} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 54 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.outputtbl4_n1 diff --git a/ql/src/test/results/clientpositive/llap/infer_bucket_sort_num_buckets.q.out b/ql/src/test/results/clientpositive/llap/infer_bucket_sort_num_buckets.q.out index 6deaafcb57..bab2ac51a9 100644 --- a/ql/src/test/results/clientpositive/llap/infer_bucket_sort_num_buckets.q.out +++ b/ql/src/test/results/clientpositive/llap/infer_bucket_sort_num_buckets.q.out @@ -80,30 +80,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -113,10 +103,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -129,30 +117,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -162,10 +140,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -272,7 +248,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -282,10 +257,8 @@ STAGE PLANS: name default.test_table_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test_table_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table_n0 TotalFiles: 1 @@ -307,7 +280,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -317,10 +289,8 @@ STAGE PLANS: name default.test_table_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test_table_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table_n0 diff --git a/ql/src/test/results/clientpositive/llap/input23.q.out b/ql/src/test/results/clientpositive/llap/input23.q.out index e0799d4d07..5600e9fbac 100644 --- a/ql/src/test/results/clientpositive/llap/input23.q.out +++ b/ql/src/test/results/clientpositive/llap/input23.q.out @@ -65,30 +65,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -98,10 +88,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/llap/input42.q.out b/ql/src/test/results/clientpositive/llap/input42.q.out index 20d204cd54..8b0ba89ab4 100644 --- a/ql/src/test/results/clientpositive/llap/input42.q.out +++ b/ql/src/test/results/clientpositive/llap/input42.q.out @@ -30,30 +30,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -63,10 +53,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -77,30 +65,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -110,10 +88,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1171,30 +1147,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1204,10 +1170,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1218,30 +1182,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1251,10 +1205,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1693,30 +1645,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1726,10 +1668,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1740,30 +1680,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1773,10 +1703,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/llap/input_part1.q.out b/ql/src/test/results/clientpositive/llap/input_part1.q.out index eb74be644c..acd0ac6af2 100644 --- a/ql/src/test/results/clientpositive/llap/input_part1.q.out +++ b/ql/src/test/results/clientpositive/llap/input_part1.q.out @@ -64,8 +64,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,hr,ds @@ -73,14 +71,8 @@ STAGE PLANS: columns.types int:string:string:string #### A masked pattern was here #### name default.dest1_n45 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n45 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n45 TotalFiles: 1 @@ -119,30 +111,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -152,10 +134,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -207,8 +187,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,hr,ds @@ -216,14 +194,8 @@ STAGE PLANS: columns.types int:string:string:string #### A masked pattern was here #### name default.dest1_n45 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n45 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n45 diff --git a/ql/src/test/results/clientpositive/llap/input_part2.q.out b/ql/src/test/results/clientpositive/llap/input_part2.q.out index 3a5dde72f8..cb411ca417 100644 --- a/ql/src/test/results/clientpositive/llap/input_part2.q.out +++ b/ql/src/test/results/clientpositive/llap/input_part2.q.out @@ -77,8 +77,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,hr,ds @@ -86,14 +84,8 @@ STAGE PLANS: columns.types int:string:string:string #### A masked pattern was here #### name default.dest1_n84 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n84 TotalFiles: 1 @@ -138,8 +130,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,hr,ds @@ -147,14 +137,8 @@ STAGE PLANS: columns.types int:string:string:string #### A masked pattern was here #### name default.dest2_n20 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest2_n20 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2_n20 TotalFiles: 1 @@ -193,30 +177,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -226,10 +200,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -242,30 +214,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -275,10 +237,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -364,8 +324,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,hr,ds @@ -373,14 +331,8 @@ STAGE PLANS: columns.types int:string:string:string #### A masked pattern was here #### name default.dest1_n84 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n84 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n84 @@ -403,8 +355,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,hr,ds @@ -412,14 +362,8 @@ STAGE PLANS: columns.types int:string:string:string #### A masked pattern was here #### name default.dest2_n20 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest2_n20 { i32 key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest2_n20 diff --git a/ql/src/test/results/clientpositive/llap/input_part7.q.out b/ql/src/test/results/clientpositive/llap/input_part7.q.out index 6734bf1554..297ff38748 100644 --- a/ql/src/test/results/clientpositive/llap/input_part7.q.out +++ b/ql/src/test/results/clientpositive/llap/input_part7.q.out @@ -78,30 +78,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -111,10 +101,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -127,30 +115,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -160,10 +138,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -212,30 +188,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -245,10 +211,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -261,30 +225,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -294,10 +248,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/llap/input_part9.q.out b/ql/src/test/results/clientpositive/llap/input_part9.q.out index ec2c8930a6..5c6449ae85 100644 --- a/ql/src/test/results/clientpositive/llap/input_part9.q.out +++ b/ql/src/test/results/clientpositive/llap/input_part9.q.out @@ -30,30 +30,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -63,10 +53,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -77,30 +65,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -110,10 +88,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/llap/join17.q.out b/ql/src/test/results/clientpositive/llap/join17.q.out index 88f4bf6d71..e432fb8019 100644 --- a/ql/src/test/results/clientpositive/llap/join17.q.out +++ b/ql/src/test/results/clientpositive/llap/join17.q.out @@ -77,30 +77,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -108,14 +98,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -158,30 +142,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -189,14 +163,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -231,8 +199,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key1":"true","key2":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,value1,key2,value2 @@ -240,14 +206,8 @@ STAGE PLANS: columns.types int:string:int:string #### A masked pattern was here #### name default.dest1_n121 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n121 { i32 key1, string value1, i32 key2, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n121 TotalFiles: 1 @@ -318,8 +278,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key1":"true","key2":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key1,value1,key2,value2 @@ -327,14 +285,8 @@ STAGE PLANS: columns.types int:string:int:string #### A masked pattern was here #### name default.dest1_n121 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n121 { i32 key1, string value1, i32 key2, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n121 diff --git a/ql/src/test/results/clientpositive/llap/join26.q.out b/ql/src/test/results/clientpositive/llap/join26.q.out index 724ec9e371..71a0af2324 100644 --- a/ql/src/test/results/clientpositive/llap/join26.q.out +++ b/ql/src/test/results/clientpositive/llap/join26.q.out @@ -103,30 +103,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -134,14 +124,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -183,30 +167,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -214,14 +188,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src1 name: default.src1 @@ -270,8 +238,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -279,14 +245,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n10 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n10 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n10 TotalFiles: 1 @@ -325,30 +285,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -358,10 +308,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -413,8 +361,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -422,14 +368,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n10 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n10 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n10 diff --git a/ql/src/test/results/clientpositive/llap/join32.q.out b/ql/src/test/results/clientpositive/llap/join32.q.out index 24b3e118dd..20db9a3bbd 100644 --- a/ql/src/test/results/clientpositive/llap/join32.q.out +++ b/ql/src/test/results/clientpositive/llap/join32.q.out @@ -103,30 +103,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -134,14 +124,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -184,30 +168,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -215,14 +189,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src1 name: default.src1 @@ -271,8 +239,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -280,14 +246,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n12 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n12 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n12 TotalFiles: 1 @@ -326,30 +286,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -359,10 +309,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -414,8 +362,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -423,14 +369,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n12 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n12 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n12 diff --git a/ql/src/test/results/clientpositive/llap/join32_lessSize.q.out b/ql/src/test/results/clientpositive/llap/join32_lessSize.q.out index c7b8bf69e4..106050f45f 100644 --- a/ql/src/test/results/clientpositive/llap/join32_lessSize.q.out +++ b/ql/src/test/results/clientpositive/llap/join32_lessSize.q.out @@ -111,30 +111,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -142,14 +132,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -192,30 +176,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -223,14 +197,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src1 name: default.src1 @@ -279,8 +247,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -288,14 +254,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n21 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n21 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n21 TotalFiles: 1 @@ -334,30 +294,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -367,10 +317,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -422,8 +370,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -431,14 +377,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n21 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n21 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n21 @@ -667,30 +607,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -698,14 +628,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src1 name: default.src1 @@ -748,30 +672,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -779,14 +693,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src1 name: default.src1 @@ -841,30 +749,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -872,14 +770,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -928,8 +820,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -937,14 +827,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n21 - numFiles 1 - numRows 85 - rawDataSize 1600 - serialization.ddl struct dest_j1_n21 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1685 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n21 TotalFiles: 1 @@ -980,30 +864,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1011,14 +885,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -1070,8 +938,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -1079,14 +945,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n21 - numFiles 1 - numRows 85 - rawDataSize 1600 - serialization.ddl struct dest_j1_n21 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1685 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n21 @@ -1311,30 +1171,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1342,14 +1192,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -1392,30 +1236,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1423,14 +1257,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src1 name: default.src1 @@ -1479,8 +1307,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -1488,14 +1314,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j2_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j2_n1 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j2_n1 TotalFiles: 1 @@ -1534,30 +1354,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1567,10 +1377,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1622,8 +1430,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -1631,14 +1437,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j2_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j2_n1 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j2_n1 @@ -1853,30 +1653,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1884,14 +1674,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src1 name: default.src1 @@ -1933,30 +1717,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1964,14 +1738,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -2020,8 +1788,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -2029,14 +1795,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j2_n1 - numFiles 1 - numRows 85 - rawDataSize 1600 - serialization.ddl struct dest_j2_n1 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1685 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j2_n1 TotalFiles: 1 @@ -2075,30 +1835,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2108,10 +1858,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2187,8 +1935,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -2196,14 +1942,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j2_n1 - numFiles 1 - numRows 85 - rawDataSize 1600 - serialization.ddl struct dest_j2_n1 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1685 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j2_n1 diff --git a/ql/src/test/results/clientpositive/llap/join33.q.out b/ql/src/test/results/clientpositive/llap/join33.q.out index 530e4a5e30..bfbb4ba310 100644 --- a/ql/src/test/results/clientpositive/llap/join33.q.out +++ b/ql/src/test/results/clientpositive/llap/join33.q.out @@ -103,30 +103,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -134,14 +124,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -184,30 +168,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -215,14 +189,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src1 name: default.src1 @@ -271,8 +239,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -280,14 +246,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n7 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n7 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n7 TotalFiles: 1 @@ -326,30 +286,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -359,10 +309,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -414,8 +362,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -423,14 +369,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n7 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n7 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n7 diff --git a/ql/src/test/results/clientpositive/llap/join34.q.out b/ql/src/test/results/clientpositive/llap/join34.q.out index 0802efe2ac..a58ee4af3d 100644 --- a/ql/src/test/results/clientpositive/llap/join34.q.out +++ b/ql/src/test/results/clientpositive/llap/join34.q.out @@ -102,8 +102,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -111,14 +109,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n1 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n1 TotalFiles: 1 @@ -154,30 +146,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -185,14 +167,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -241,8 +217,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -250,14 +224,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n1 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n1 TotalFiles: 1 @@ -293,30 +261,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -324,14 +282,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -385,30 +337,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -416,14 +358,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src1 name: default.src1 @@ -477,8 +413,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -486,14 +420,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.dest_j1_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n1 { string key, string value, string val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n1 diff --git a/ql/src/test/results/clientpositive/llap/join35.q.out b/ql/src/test/results/clientpositive/llap/join35.q.out index fb4a512e0d..cd435a606c 100644 --- a/ql/src/test/results/clientpositive/llap/join35.q.out +++ b/ql/src/test/results/clientpositive/llap/join35.q.out @@ -101,30 +101,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -132,14 +122,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -185,30 +169,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -216,14 +190,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -277,30 +245,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -308,14 +266,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src1 name: default.src1 @@ -359,8 +311,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -368,14 +318,8 @@ STAGE PLANS: columns.types string:string:int #### A masked pattern was here #### name default.dest_j1_n24 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n24 { string key, string value, i32 val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n24 TotalFiles: 1 @@ -471,8 +415,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -480,14 +422,8 @@ STAGE PLANS: columns.types string:string:int #### A masked pattern was here #### name default.dest_j1_n24 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n24 { string key, string value, i32 val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n24 TotalFiles: 1 @@ -527,8 +463,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","val2":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,val2 @@ -536,14 +470,8 @@ STAGE PLANS: columns.types string:string:int #### A masked pattern was here #### name default.dest_j1_n24 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest_j1_n24 { string key, string value, i32 val2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest_j1_n24 diff --git a/ql/src/test/results/clientpositive/llap/join9.q.out b/ql/src/test/results/clientpositive/llap/join9.q.out index d28d5fe047..2ce0126eb0 100644 --- a/ql/src/test/results/clientpositive/llap/join9.q.out +++ b/ql/src/test/results/clientpositive/llap/join9.q.out @@ -83,30 +83,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -116,10 +106,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -162,30 +150,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -193,14 +171,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -235,8 +207,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -244,14 +214,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.dest1_n39 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n39 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n39 TotalFiles: 1 @@ -322,8 +286,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -331,14 +293,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.dest1_n39 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n39 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n39 diff --git a/ql/src/test/results/clientpositive/llap/join_filters_overlap.q.out b/ql/src/test/results/clientpositive/llap/join_filters_overlap.q.out index dd0f0337e9..cb87a53bb3 100644 --- a/ql/src/test/results/clientpositive/llap/join_filters_overlap.q.out +++ b/ql/src/test/results/clientpositive/llap/join_filters_overlap.q.out @@ -91,30 +91,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -122,14 +112,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.a_n4 name: default.a_n4 @@ -172,30 +156,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -203,14 +177,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.a_n4 name: default.a_n4 @@ -406,30 +374,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -437,14 +395,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.a_n4 name: default.a_n4 @@ -482,30 +434,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -513,14 +455,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.a_n4 name: default.a_n4 @@ -716,30 +652,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -747,14 +673,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.a_n4 name: default.a_n4 @@ -792,30 +712,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -823,14 +733,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.a_n4 name: default.a_n4 @@ -1043,30 +947,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1074,14 +968,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.a_n4 name: default.a_n4 @@ -1119,30 +1007,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1150,14 +1028,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.a_n4 name: default.a_n4 @@ -1392,30 +1264,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1423,14 +1285,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.a_n4 name: default.a_n4 @@ -1473,30 +1329,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1504,14 +1350,8 @@ STAGE PLANS: columns.types int:int #### A masked pattern was here #### name default.a_n4 - numFiles 1 - numRows 3 - rawDataSize 18 - serialization.ddl struct a_n4 { i32 key, i32 value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 21 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.a_n4 name: default.a_n4 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_1.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_1.q.out index 189335a3ad..fa0d8f0ed8 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_1.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_1.q.out @@ -66,7 +66,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -76,10 +75,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_dynamic_part_n0 TotalFiles: 1 @@ -121,30 +118,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -154,10 +141,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -170,30 +155,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -203,10 +178,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -267,7 +240,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -277,10 +249,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_dynamic_part_n0 @@ -456,30 +426,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part_n0 - numFiles 2 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct list_bucketing_dynamic_part_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -489,10 +449,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_dynamic_part_n0 name: default.list_bucketing_dynamic_part_n0 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_10.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_10.q.out index 3ce63ffa0a..a22819003b 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_10.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_10.q.out @@ -65,7 +65,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -75,10 +74,8 @@ STAGE PLANS: name default.list_bucketing_static_part partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part TotalFiles: 1 @@ -117,30 +114,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -148,14 +135,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -215,7 +196,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -225,10 +205,8 @@ STAGE PLANS: name default.list_bucketing_static_part partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_11.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_11.q.out index 3cfe511333..4fe1310685 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_11.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_11.q.out @@ -65,7 +65,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -75,10 +74,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n3 TotalFiles: 1 @@ -117,30 +114,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -148,14 +135,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -215,7 +196,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -225,10 +205,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n3 @@ -334,30 +312,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_static_part_n3 - numFiles 4 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 4812 - serialization.ddl struct list_bucketing_static_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 5522 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -367,10 +335,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n3 name: default.list_bucketing_static_part_n3 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_12.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_12.q.out index 297dd32b35..9b2ec89b48 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_12.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_12.q.out @@ -65,7 +65,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns col1,col2,col3,col4,col5 @@ -75,10 +74,8 @@ STAGE PLANS: name default.list_bucketing_mul_col_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_mul_col_n0 { string col1, string col2, string col3, string col4, string col5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_mul_col_n0 TotalFiles: 1 @@ -117,30 +114,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -148,14 +135,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -215,7 +196,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns col1,col2,col3,col4,col5 @@ -225,10 +205,8 @@ STAGE PLANS: name default.list_bucketing_mul_col_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_mul_col_n0 { string col1, string col2, string col3, string col4, string col5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_mul_col_n0 @@ -342,30 +320,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col1":"true","col2":"true","col3":"true","col4":"true","col5":"true"}} - bucket_count -1 column.name.delimiter , columns col1,col2,col3,col4,col5 - columns.comments columns.types string:string:string:string:string #### A masked pattern was here #### name default.list_bucketing_mul_col_n0 - numFiles 4 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 6312 - serialization.ddl struct list_bucketing_mul_col_n0 { string col1, string col2, string col3, string col4, string col5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 7094 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns col1,col2,col3,col4,col5 @@ -375,10 +343,8 @@ STAGE PLANS: name default.list_bucketing_mul_col_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_mul_col_n0 { string col1, string col2, string col3, string col4, string col5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_mul_col_n0 name: default.list_bucketing_mul_col_n0 @@ -442,30 +408,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col1":"true","col2":"true","col3":"true","col4":"true","col5":"true"}} - bucket_count -1 column.name.delimiter , columns col1,col2,col3,col4,col5 - columns.comments columns.types string:string:string:string:string #### A masked pattern was here #### name default.list_bucketing_mul_col_n0 - numFiles 4 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 6312 - serialization.ddl struct list_bucketing_mul_col_n0 { string col1, string col2, string col3, string col4, string col5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 7094 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns col1,col2,col3,col4,col5 @@ -475,10 +431,8 @@ STAGE PLANS: name default.list_bucketing_mul_col_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_mul_col_n0 { string col1, string col2, string col3, string col4, string col5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_mul_col_n0 name: default.list_bucketing_mul_col_n0 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_13.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_13.q.out index e272f2180c..ac035e3203 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_13.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_13.q.out @@ -65,7 +65,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns col1,col2,col3,col4,col5 @@ -75,10 +74,8 @@ STAGE PLANS: name default.list_bucketing_mul_col partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_mul_col { string col1, string col2, string col3, string col4, string col5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_mul_col TotalFiles: 1 @@ -117,30 +114,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -148,14 +135,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -215,7 +196,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns col1,col2,col3,col4,col5 @@ -225,10 +205,8 @@ STAGE PLANS: name default.list_bucketing_mul_col partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_mul_col { string col1, string col2, string col3, string col4, string col5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_mul_col @@ -342,30 +320,20 @@ STAGE PLANS: ds 2008-04-08 hr 2013-01-23+18:00:99 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col1":"true","col2":"true","col3":"true","col4":"true","col5":"true"}} - bucket_count -1 column.name.delimiter , columns col1,col2,col3,col4,col5 - columns.comments columns.types string:string:string:string:string #### A masked pattern was here #### name default.list_bucketing_mul_col - numFiles 4 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 6312 - serialization.ddl struct list_bucketing_mul_col { string col1, string col2, string col3, string col4, string col5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 7094 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns col1,col2,col3,col4,col5 @@ -375,10 +343,8 @@ STAGE PLANS: name default.list_bucketing_mul_col partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_mul_col { string col1, string col2, string col3, string col4, string col5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_mul_col name: default.list_bucketing_mul_col diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_14.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_14.q.out index 681652d8cd..e79a6e7a58 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_14.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_14.q.out @@ -58,8 +58,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -67,14 +65,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.list_bucketing - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct list_bucketing { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing TotalFiles: 1 @@ -110,30 +102,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -141,14 +123,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -200,8 +176,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -209,14 +183,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.list_bucketing - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct list_bucketing { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_2.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_2.q.out index bd1df7c92d..dcae664bb5 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_2.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_2.q.out @@ -71,7 +71,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -81,10 +80,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n4 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n4 TotalFiles: 1 @@ -126,30 +123,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -159,10 +146,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -175,30 +160,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -208,10 +183,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -272,7 +245,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -282,10 +254,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n4 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n4 @@ -413,30 +383,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_static_part_n4 - numFiles 3 - numRows 1000 partition_columns ds/hr partition_columns.types string:string - rawDataSize 9624 - serialization.ddl struct list_bucketing_static_part_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 10659 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -446,10 +406,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n4 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n4 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n4 name: default.list_bucketing_static_part_n4 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_3.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_3.q.out index ad1b340a9d..a36d0b8e50 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_3.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_3.q.out @@ -61,7 +61,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -71,10 +70,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n1 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_static_part_n1 TotalFiles: 1 @@ -116,30 +113,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -149,10 +136,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -165,30 +150,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -198,10 +173,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -262,7 +235,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -272,10 +244,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n1 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_static_part_n1 @@ -405,30 +375,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_static_part_n1 - numFiles 2 - numRows 1000 partition_columns ds/hr partition_columns.types string:string - rawDataSize 10624 - serialization.ddl struct list_bucketing_static_part_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -438,10 +398,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n1 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_static_part_n1 name: default.list_bucketing_static_part_n1 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_4.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_4.q.out index 25a29beee5..68144f03f1 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_4.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_4.q.out @@ -71,7 +71,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -81,10 +80,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n2 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n2 TotalFiles: 1 @@ -126,30 +123,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -159,10 +146,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -175,30 +160,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -208,10 +183,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -272,7 +245,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -282,10 +254,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n2 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n2 @@ -422,7 +392,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -432,10 +401,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n2 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n2 TotalFiles: 1 @@ -477,30 +444,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -510,10 +467,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -526,30 +481,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -559,10 +504,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -623,7 +566,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -633,10 +575,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n2 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n2 @@ -764,30 +704,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_static_part_n2 - numFiles 3 - numRows 1000 partition_columns ds/hr partition_columns.types string:string - rawDataSize 9624 - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 10659 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -797,10 +727,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n2 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n2 name: default.list_bucketing_static_part_n2 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_5.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_5.q.out index 7aa4b23f1f..e5cb4f95e6 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_5.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_5.q.out @@ -66,7 +66,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -76,10 +75,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n1 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_dynamic_part_n1 TotalFiles: 1 @@ -121,30 +118,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -154,10 +141,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -170,30 +155,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -203,10 +178,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -267,7 +240,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -277,10 +249,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n1 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_dynamic_part_n1 @@ -463,30 +433,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part_n1 - numFiles 3 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct list_bucketing_dynamic_part_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -496,10 +456,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n1 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_dynamic_part_n1 name: default.list_bucketing_dynamic_part_n1 @@ -510,30 +468,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part_n1 - numFiles 3 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct list_bucketing_dynamic_part_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -543,10 +491,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n1 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.list_bucketing_dynamic_part_n1 name: default.list_bucketing_dynamic_part_n1 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_6.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_6.q.out index a3dccdfe80..dd1e97bec2 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_6.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_6.q.out @@ -70,7 +70,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -80,10 +79,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n3 TotalFiles: 1 @@ -125,30 +122,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -158,10 +145,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -174,30 +159,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -207,10 +182,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -271,7 +244,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -281,10 +253,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n3 @@ -465,7 +435,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -475,10 +444,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n3 TotalFiles: 1 @@ -520,30 +487,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -553,10 +510,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -569,30 +524,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -602,10 +547,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -666,7 +609,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -676,10 +618,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n3 @@ -854,30 +794,20 @@ STAGE PLANS: ds 2008-04-08 hr a1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part_n3 - numFiles 1 - numRows 16 partition_columns ds/hr partition_columns.types string:string - rawDataSize 136 - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 235 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -887,10 +817,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n3 name: default.list_bucketing_dynamic_part_n3 @@ -901,30 +829,20 @@ STAGE PLANS: ds 2008-04-08 hr b1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part_n3 - numFiles 3 - numRows 984 partition_columns ds/hr partition_columns.types string:string - rawDataSize 9488 - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 10495 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -934,10 +852,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part_n3 name: default.list_bucketing_dynamic_part_n3 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_7.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_7.q.out index b517fa046f..87cb08fe12 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_7.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_7.q.out @@ -70,7 +70,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -80,10 +79,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part TotalFiles: 1 @@ -125,30 +122,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -158,10 +145,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -174,30 +159,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -207,10 +182,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -271,7 +244,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -281,10 +253,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part @@ -465,7 +435,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -475,10 +444,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part TotalFiles: 1 @@ -520,30 +487,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -553,10 +510,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -569,30 +524,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -602,10 +547,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -666,7 +609,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -676,10 +618,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part @@ -854,30 +794,20 @@ STAGE PLANS: ds 2008-04-08 hr a1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part - numFiles 1 - numRows 16 partition_columns ds/hr partition_columns.types string:string - rawDataSize 136 - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 235 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -887,10 +817,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part name: default.list_bucketing_dynamic_part @@ -901,30 +829,20 @@ STAGE PLANS: ds 2008-04-08 hr b1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_dynamic_part - numFiles 2 - numRows 984 partition_columns ds/hr partition_columns.types string:string - rawDataSize 9488 - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 10416 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -934,10 +852,8 @@ STAGE PLANS: name default.list_bucketing_dynamic_part partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_dynamic_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_dynamic_part name: default.list_bucketing_dynamic_part diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_dml_9.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_dml_9.q.out index 5e6e9cc935..1938bfbf4e 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_dml_9.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_dml_9.q.out @@ -71,7 +71,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -81,10 +80,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n0 TotalFiles: 1 @@ -126,30 +123,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -159,10 +146,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -175,30 +160,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -208,10 +183,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -272,7 +245,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -282,10 +254,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n0 @@ -422,7 +392,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -432,10 +401,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n0 TotalFiles: 1 @@ -477,30 +444,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -510,10 +467,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -526,30 +481,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -559,10 +504,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -623,7 +566,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -633,10 +575,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n0 @@ -764,30 +704,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.list_bucketing_static_part_n0 - numFiles 3 - numRows 1000 partition_columns ds/hr partition_columns.types string:string - rawDataSize 9624 - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 10659 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -797,10 +727,8 @@ STAGE PLANS: name default.list_bucketing_static_part_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct list_bucketing_static_part_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.list_bucketing_static_part_n0 name: default.list_bucketing_static_part_n0 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_1.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_1.q.out index 441b71d52c..9922619be9 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_1.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_1.q.out @@ -104,30 +104,20 @@ STAGE PLANS: ds 1 hr 4 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.fact_daily - numFiles 3 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -137,10 +127,8 @@ STAGE PLANS: name default.fact_daily partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct fact_daily { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily name: default.fact_daily @@ -196,30 +184,20 @@ STAGE PLANS: ds 1 hr 4 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.fact_daily - numFiles 3 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -229,10 +207,8 @@ STAGE PLANS: name default.fact_daily partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct fact_daily { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily name: default.fact_daily @@ -289,30 +265,20 @@ STAGE PLANS: ds 1 hr 4 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.fact_daily - numFiles 3 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -322,10 +288,8 @@ STAGE PLANS: name default.fact_daily partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct fact_daily { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily name: default.fact_daily @@ -380,30 +344,20 @@ STAGE PLANS: ds 1 hr 4 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.fact_daily - numFiles 3 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -413,10 +367,8 @@ STAGE PLANS: name default.fact_daily partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct fact_daily { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily name: default.fact_daily diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_2.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_2.q.out index 4ef05435b7..66be4d091d 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_2.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_2.q.out @@ -104,30 +104,20 @@ STAGE PLANS: ds 1 hr 4 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.fact_daily_n2 - numFiles 3 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -137,10 +127,8 @@ STAGE PLANS: name default.fact_daily_n2 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct fact_daily_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n2 name: default.fact_daily_n2 @@ -196,30 +184,20 @@ STAGE PLANS: ds 1 hr 4 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.fact_daily_n2 - numFiles 3 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -229,10 +207,8 @@ STAGE PLANS: name default.fact_daily_n2 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct fact_daily_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n2 name: default.fact_daily_n2 @@ -291,30 +267,20 @@ STAGE PLANS: ds 1 hr 4 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.fact_daily_n2 - numFiles 3 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -324,10 +290,8 @@ STAGE PLANS: name default.fact_daily_n2 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct fact_daily_n2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n2 name: default.fact_daily_n2 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_3.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_3.q.out index 8f3dd7dc25..bc051ea5bd 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_3.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_query_multiskew_3.q.out @@ -214,30 +214,20 @@ STAGE PLANS: ds 1 hr 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.fact_daily_n3 - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -247,10 +237,8 @@ STAGE PLANS: name default.fact_daily_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct fact_daily_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n3 name: default.fact_daily_n3 @@ -340,30 +328,20 @@ STAGE PLANS: ds 1 hr 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.fact_daily_n3 - numFiles 3 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -373,10 +351,8 @@ STAGE PLANS: name default.fact_daily_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct fact_daily_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n3 name: default.fact_daily_n3 @@ -434,30 +410,20 @@ STAGE PLANS: ds 1 hr 3 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.fact_daily_n3 - numFiles 2 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct fact_daily_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -467,10 +433,8 @@ STAGE PLANS: name default.fact_daily_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct fact_daily_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n3 name: default.fact_daily_n3 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_1.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_1.q.out index 1bc4d62e5d..564f306a1f 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_1.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_1.q.out @@ -155,28 +155,21 @@ STAGE PLANS: partition values: ds 1 properties: - bucket_count -1 column.name.delimiter , columns x - columns.comments columns.types int #### A masked pattern was here #### name default.fact_daily_n4 - numFiles 2 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n4 { i32 x} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 8 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: EXTERNAL TRUE - bucket_count -1 bucketing_version 2 column.name.delimiter , columns x @@ -186,10 +179,8 @@ STAGE PLANS: name default.fact_daily_n4 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n4 { i32 x} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n4 name: default.fact_daily_n4 @@ -244,28 +235,21 @@ STAGE PLANS: partition values: ds 1 properties: - bucket_count -1 column.name.delimiter , columns x - columns.comments columns.types int #### A masked pattern was here #### name default.fact_daily_n4 - numFiles 2 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n4 { i32 x} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 8 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: EXTERNAL TRUE - bucket_count -1 bucketing_version 2 column.name.delimiter , columns x @@ -275,10 +259,8 @@ STAGE PLANS: name default.fact_daily_n4 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n4 { i32 x} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n4 name: default.fact_daily_n4 @@ -333,28 +315,21 @@ STAGE PLANS: partition values: ds 1 properties: - bucket_count -1 column.name.delimiter , columns x - columns.comments columns.types int #### A masked pattern was here #### name default.fact_daily_n4 - numFiles 2 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n4 { i32 x} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 8 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: EXTERNAL TRUE - bucket_count -1 bucketing_version 2 column.name.delimiter , columns x @@ -364,10 +339,8 @@ STAGE PLANS: name default.fact_daily_n4 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n4 { i32 x} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n4 name: default.fact_daily_n4 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_2.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_2.q.out index a60bfe65f8..f45e9d470e 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_2.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_2.q.out @@ -158,28 +158,21 @@ STAGE PLANS: partition values: ds 1 properties: - bucket_count -1 column.name.delimiter , columns x,y - columns.comments columns.types int:string #### A masked pattern was here #### name default.fact_daily_n5 - numFiles 2 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n5 { i32 x, string y} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 24 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: EXTERNAL TRUE - bucket_count -1 bucketing_version 2 column.name.delimiter , columns x,y @@ -189,10 +182,8 @@ STAGE PLANS: name default.fact_daily_n5 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n5 { i32 x, string y} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n5 name: default.fact_daily_n5 @@ -247,28 +238,21 @@ STAGE PLANS: partition values: ds 1 properties: - bucket_count -1 column.name.delimiter , columns x,y - columns.comments columns.types int:string #### A masked pattern was here #### name default.fact_daily_n5 - numFiles 2 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n5 { i32 x, string y} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 24 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: EXTERNAL TRUE - bucket_count -1 bucketing_version 2 column.name.delimiter , columns x,y @@ -278,10 +262,8 @@ STAGE PLANS: name default.fact_daily_n5 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n5 { i32 x, string y} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n5 name: default.fact_daily_n5 @@ -381,28 +363,21 @@ STAGE PLANS: partition values: ds 1 properties: - bucket_count -1 column.name.delimiter , columns x,y - columns.comments columns.types int:string #### A masked pattern was here #### name default.fact_daily_n5 - numFiles 2 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n5 { i32 x, string y} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 24 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: EXTERNAL TRUE - bucket_count -1 bucketing_version 2 column.name.delimiter , columns x,y @@ -412,10 +387,8 @@ STAGE PLANS: name default.fact_daily_n5 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n5 { i32 x, string y} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n5 name: default.fact_daily_n5 @@ -543,28 +516,21 @@ STAGE PLANS: partition values: ds 1 properties: - bucket_count -1 column.name.delimiter , columns x,y - columns.comments columns.types int:string #### A masked pattern was here #### name default.fact_daily_n5 - numFiles 2 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n5 { i32 x, string y} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 24 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: EXTERNAL TRUE - bucket_count -1 bucketing_version 2 column.name.delimiter , columns x,y @@ -574,10 +540,8 @@ STAGE PLANS: name default.fact_daily_n5 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n5 { i32 x, string y} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n5 name: default.fact_daily_n5 diff --git a/ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_3.q.out b/ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_3.q.out index 1bfbfd3f65..106050b5b9 100644 --- a/ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_3.q.out +++ b/ql/src/test/results/clientpositive/llap/list_bucket_query_oneskew_3.q.out @@ -178,28 +178,21 @@ STAGE PLANS: partition values: ds 1 properties: - bucket_count -1 column.name.delimiter , columns x,y,z - columns.comments columns.types int:string:string #### A masked pattern was here #### name default.fact_daily_n0 - numFiles 3 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n0 { i32 x, string y, string z} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 117 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: EXTERNAL TRUE - bucket_count -1 bucketing_version 2 column.name.delimiter , columns x,y,z @@ -209,10 +202,8 @@ STAGE PLANS: name default.fact_daily_n0 partition_columns ds partition_columns.types string - serialization.ddl struct fact_daily_n0 { i32 x, string y, string z} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.fact_daily_n0 name: default.fact_daily_n0 diff --git a/ql/src/test/results/clientpositive/llap/llap_nullscan.q.out b/ql/src/test/results/clientpositive/llap/llap_nullscan.q.out index 07a30ce067..3adecd3ce3 100644 --- a/ql/src/test/results/clientpositive/llap/llap_nullscan.q.out +++ b/ql/src/test/results/clientpositive/llap/llap_nullscan.q.out @@ -91,31 +91,21 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,ds,hr - columns.comments columns.types string:string:string:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src_orc_n1 - numFiles 1 - numRows 10 - rawDataSize 3560 - serialization.ddl struct src_orc_n1 { string key, string value, string ds, string hr} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 646 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,ds,hr @@ -124,14 +114,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src_orc_n1 - numFiles 1 - numRows 10 - rawDataSize 3560 - serialization.ddl struct src_orc_n1 { string key, string value, string ds, string hr} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 646 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.src_orc_n1 name: default.src_orc_n1 diff --git a/ql/src/test/results/clientpositive/llap/load_dyn_part8.q.out b/ql/src/test/results/clientpositive/llap/load_dyn_part8.q.out index 28e12fd289..f46a6bcce2 100644 --- a/ql/src/test/results/clientpositive/llap/load_dyn_part8.q.out +++ b/ql/src/test/results/clientpositive/llap/load_dyn_part8.q.out @@ -98,7 +98,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -108,10 +107,8 @@ STAGE PLANS: name default.nzhang_part8_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct nzhang_part8_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part8_n0 TotalFiles: 1 @@ -160,7 +157,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -170,10 +166,8 @@ STAGE PLANS: name default.nzhang_part8_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct nzhang_part8_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part8_n0 TotalFiles: 1 @@ -215,30 +209,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -248,10 +232,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -264,30 +246,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -297,10 +269,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -313,30 +283,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -346,10 +306,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -362,30 +320,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -395,10 +343,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -499,7 +445,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -509,10 +454,8 @@ STAGE PLANS: name default.nzhang_part8_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct nzhang_part8_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part8_n0 @@ -533,7 +476,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -543,10 +485,8 @@ STAGE PLANS: name default.nzhang_part8_n0 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct nzhang_part8_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.nzhang_part8_n0 diff --git a/ql/src/test/results/clientpositive/llap/louter_join_ppr.q.out b/ql/src/test/results/clientpositive/llap/louter_join_ppr.q.out index f39ba92a88..ac36d6c20f 100644 --- a/ql/src/test/results/clientpositive/llap/louter_join_ppr.q.out +++ b/ql/src/test/results/clientpositive/llap/louter_join_ppr.q.out @@ -82,30 +82,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -113,14 +103,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -166,30 +150,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -199,10 +173,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -215,30 +187,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -248,10 +210,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -427,30 +387,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -460,10 +410,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -476,30 +424,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -509,10 +447,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -556,30 +492,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -587,14 +513,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -766,30 +686,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -797,14 +707,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -850,30 +754,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -883,10 +777,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -899,30 +791,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -932,10 +814,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1111,30 +991,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1144,10 +1014,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1160,30 +1028,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1193,10 +1051,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1240,30 +1096,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1271,14 +1117,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/llap/mapjoin_mapjoin.q.out b/ql/src/test/results/clientpositive/llap/mapjoin_mapjoin.q.out index f69a950361..405b71e217 100644 --- a/ql/src/test/results/clientpositive/llap/mapjoin_mapjoin.q.out +++ b/ql/src/test/results/clientpositive/llap/mapjoin_mapjoin.q.out @@ -117,30 +117,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -150,10 +140,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -166,30 +154,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -199,10 +177,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -215,30 +191,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -248,10 +214,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -264,30 +228,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -297,10 +251,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -345,30 +297,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -376,14 +318,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src1 - numFiles 1 - numRows 25 - rawDataSize 191 - serialization.ddl struct src1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src1 name: default.src1 @@ -425,30 +361,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -456,14 +382,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/llap/merge3.q.out b/ql/src/test/results/clientpositive/llap/merge3.q.out index e985f54235..141e40891e 100644 --- a/ql/src/test/results/clientpositive/llap/merge3.q.out +++ b/ql/src/test/results/clientpositive/llap/merge3.q.out @@ -143,30 +143,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.merge_src - numFiles 1 - numRows 2000 - rawDataSize 21248 - serialization.ddl struct merge_src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 23248 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -174,14 +164,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.merge_src - numFiles 1 - numRows 2000 - rawDataSize 21248 - serialization.ddl struct merge_src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 23248 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.merge_src name: default.merge_src @@ -2370,7 +2354,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2380,10 +2363,8 @@ STAGE PLANS: name default.merge_src_part2 partition_columns ds partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.merge_src_part2 TotalFiles: 1 @@ -2424,30 +2405,20 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.merge_src_part - numFiles 1 - numRows 1000 partition_columns ds partition_columns.types string - rawDataSize 10624 - serialization.ddl struct merge_src_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2457,10 +2428,8 @@ STAGE PLANS: name default.merge_src_part partition_columns ds partition_columns.types string - serialization.ddl struct merge_src_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.merge_src_part name: default.merge_src_part @@ -2472,30 +2441,20 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.merge_src_part - numFiles 1 - numRows 1000 partition_columns ds partition_columns.types string - rawDataSize 10624 - serialization.ddl struct merge_src_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2505,10 +2464,8 @@ STAGE PLANS: name default.merge_src_part partition_columns ds partition_columns.types string - serialization.ddl struct merge_src_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.merge_src_part name: default.merge_src_part @@ -2568,7 +2525,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2578,10 +2534,8 @@ STAGE PLANS: name default.merge_src_part2 partition_columns ds partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.merge_src_part2 @@ -4718,30 +4672,20 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.merge_src_part - numFiles 1 - numRows 1000 partition_columns ds partition_columns.types string - rawDataSize 10624 - serialization.ddl struct merge_src_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4751,10 +4695,8 @@ STAGE PLANS: name default.merge_src_part partition_columns ds partition_columns.types string - serialization.ddl struct merge_src_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.merge_src_part name: default.merge_src_part @@ -4766,30 +4708,20 @@ STAGE PLANS: partition values: ds 2008-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.merge_src_part - numFiles 1 - numRows 1000 partition_columns ds partition_columns.types string - rawDataSize 10624 - serialization.ddl struct merge_src_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4799,10 +4731,8 @@ STAGE PLANS: name default.merge_src_part partition_columns ds partition_columns.types string - serialization.ddl struct merge_src_part { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.merge_src_part name: default.merge_src_part @@ -4829,7 +4759,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4839,10 +4768,8 @@ STAGE PLANS: name default.merge_src_part2 partition_columns ds partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.merge_src_part2 TotalFiles: 1 @@ -4901,7 +4828,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4911,10 +4837,8 @@ STAGE PLANS: name default.merge_src_part2 partition_columns ds partition_columns.types string - serialization.ddl struct merge_src_part2 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.merge_src_part2 diff --git a/ql/src/test/results/clientpositive/llap/metadataonly1.q.out b/ql/src/test/results/clientpositive/llap/metadataonly1.q.out index 38a244dcce..cba437bdaf 100644 --- a/ql/src/test/results/clientpositive/llap/metadataonly1.q.out +++ b/ql/src/test/results/clientpositive/llap/metadataonly1.q.out @@ -172,30 +172,20 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test1_n12 - numFiles 0 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -205,10 +195,8 @@ STAGE PLANS: name default.test1_n12 partition_columns ds partition_columns.types string - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n12 name: default.test1_n12 @@ -327,30 +315,20 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test1_n12 - numFiles 0 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -360,10 +338,8 @@ STAGE PLANS: name default.test1_n12 partition_columns ds partition_columns.types string - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n12 name: default.test1_n12 @@ -482,30 +458,20 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test1_n12 - numFiles 0 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -515,10 +481,8 @@ STAGE PLANS: name default.test1_n12 partition_columns ds partition_columns.types string - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n12 name: default.test1_n12 @@ -652,30 +616,20 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test1_n12 - numFiles 0 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -685,10 +639,8 @@ STAGE PLANS: name default.test1_n12 partition_columns ds partition_columns.types string - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n12 name: default.test1_n12 @@ -700,30 +652,20 @@ STAGE PLANS: partition values: ds 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test1_n12 - numFiles 0 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -733,10 +675,8 @@ STAGE PLANS: name default.test1_n12 partition_columns ds partition_columns.types string - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n12 name: default.test1_n12 @@ -781,30 +721,20 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test1_n12 - numFiles 0 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -814,10 +744,8 @@ STAGE PLANS: name default.test1_n12 partition_columns ds partition_columns.types string - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n12 name: default.test1_n12 @@ -829,30 +757,20 @@ STAGE PLANS: partition values: ds 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test1_n12 - numFiles 0 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -862,10 +780,8 @@ STAGE PLANS: name default.test1_n12 partition_columns ds partition_columns.types string - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n12 name: default.test1_n12 @@ -1091,30 +1007,20 @@ STAGE PLANS: ds 1 hr 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test2_n8 - numFiles 0 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -1124,10 +1030,8 @@ STAGE PLANS: name default.test2_n8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n8 name: default.test2_n8 @@ -1139,30 +1043,20 @@ STAGE PLANS: ds 1 hr 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test2_n8 - numFiles 0 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -1172,10 +1066,8 @@ STAGE PLANS: name default.test2_n8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n8 name: default.test2_n8 @@ -1187,30 +1079,20 @@ STAGE PLANS: ds 1 hr 3 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test2_n8 - numFiles 0 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -1220,10 +1102,8 @@ STAGE PLANS: name default.test2_n8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n8 name: default.test2_n8 @@ -1357,30 +1237,20 @@ STAGE PLANS: ds 1 hr 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test2_n8 - numFiles 0 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -1390,10 +1260,8 @@ STAGE PLANS: name default.test2_n8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n8 name: default.test2_n8 @@ -1406,30 +1274,20 @@ STAGE PLANS: ds 1 hr 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test2_n8 - numFiles 0 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -1439,10 +1297,8 @@ STAGE PLANS: name default.test2_n8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n8 name: default.test2_n8 @@ -1455,30 +1311,20 @@ STAGE PLANS: ds 1 hr 3 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test2_n8 - numFiles 0 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -1488,10 +1334,8 @@ STAGE PLANS: name default.test2_n8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n8 name: default.test2_n8 @@ -1618,30 +1462,20 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test1_n12 - numFiles 0 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -1651,10 +1485,8 @@ STAGE PLANS: name default.test1_n12 partition_columns ds partition_columns.types string - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n12 name: default.test1_n12 @@ -1665,30 +1497,20 @@ STAGE PLANS: partition values: ds 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test1_n12 - numFiles 0 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -1698,10 +1520,8 @@ STAGE PLANS: name default.test1_n12 partition_columns ds partition_columns.types string - serialization.ddl struct test1_n12 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test1_n12 name: default.test1_n12 @@ -1887,30 +1707,20 @@ STAGE PLANS: ds 01:10:10 hr 01 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test2_n8 - numFiles 0 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -1920,10 +1730,8 @@ STAGE PLANS: name default.test2_n8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n8 name: default.test2_n8 @@ -1935,30 +1743,20 @@ STAGE PLANS: ds 01:10:20 hr 02 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test2_n8 - numFiles 0 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -1968,10 +1766,8 @@ STAGE PLANS: name default.test2_n8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n8 name: default.test2_n8 @@ -1983,30 +1779,20 @@ STAGE PLANS: ds 1 hr 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test2_n8 - numFiles 0 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -2016,10 +1802,8 @@ STAGE PLANS: name default.test2_n8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n8 name: default.test2_n8 @@ -2031,30 +1815,20 @@ STAGE PLANS: ds 1 hr 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test2_n8 - numFiles 0 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -2064,10 +1838,8 @@ STAGE PLANS: name default.test2_n8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n8 name: default.test2_n8 @@ -2079,30 +1851,20 @@ STAGE PLANS: ds 1 hr 3 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"a":"true","b":"true"}} - bucket_count -1 column.name.delimiter , columns a,b - columns.comments columns.types int:double #### A masked pattern was here #### name default.test2_n8 - numFiles 0 - numRows 0 partition_columns ds/hr partition_columns.types string:string - rawDataSize 0 - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns a,b @@ -2112,10 +1874,8 @@ STAGE PLANS: name default.test2_n8 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct test2_n8 { i32 a, double b} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test2_n8 name: default.test2_n8 diff --git a/ql/src/test/results/clientpositive/llap/murmur_hash_migration.q.out b/ql/src/test/results/clientpositive/llap/murmur_hash_migration.q.out index 10de2898ac..54600bc17f 100644 --- a/ql/src/test/results/clientpositive/llap/murmur_hash_migration.q.out +++ b/ql/src/test/results/clientpositive/llap/murmur_hash_migration.q.out @@ -208,25 +208,17 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n20 - numFiles 4 - numRows 150 partition_columns ds partition_columns.types string - rawDataSize 1602 - serialization.ddl struct srcbucket_mapjoin_part_n20 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1752 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -243,10 +235,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n20 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n20 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n20 name: default.srcbucket_mapjoin_part_n20 @@ -284,10 +274,8 @@ STAGE PLANS: name default.tab_part_n11 partition_columns ds partition_columns.types string - serialization.ddl struct tab_part_n11 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tab_part_n11 TotalFiles: 1 @@ -319,10 +307,8 @@ STAGE PLANS: name default.tab_part_n11 partition_columns ds partition_columns.types string - serialization.ddl struct tab_part_n11 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tab_part_n11 @@ -417,25 +403,17 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n18 - numFiles 2 - numRows 150 partition_columns ds partition_columns.types string - rawDataSize 1598 - serialization.ddl struct srcbucket_mapjoin_n18 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1748 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -452,10 +430,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_n18 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_n18 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_n18 name: default.srcbucket_mapjoin_n18 @@ -493,10 +469,8 @@ STAGE PLANS: name default.tab_n10 partition_columns ds partition_columns.types string - serialization.ddl struct tab_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tab_n10 TotalFiles: 1 @@ -528,10 +502,8 @@ STAGE PLANS: name default.tab_n10 partition_columns ds partition_columns.types string - serialization.ddl struct tab_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tab_n10 @@ -669,25 +641,17 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n18 - numFiles 2 - numRows 150 partition_columns ds partition_columns.types string - rawDataSize 1598 - serialization.ddl struct srcbucket_mapjoin_n18 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1748 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -704,10 +668,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_n18 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_n18 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_n18 name: default.srcbucket_mapjoin_n18 @@ -752,25 +714,17 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n20 - numFiles 4 - numRows 150 partition_columns ds partition_columns.types string - rawDataSize 1602 - serialization.ddl struct srcbucket_mapjoin_part_n20 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1752 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -787,10 +741,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n20 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n20 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n20 name: default.srcbucket_mapjoin_part_n20 @@ -963,25 +915,17 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.tab_part_n11 - numFiles 4 - numRows 150 partition_columns ds partition_columns.types string - rawDataSize 1602 - serialization.ddl struct tab_part_n11 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1752 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -998,10 +942,8 @@ STAGE PLANS: name default.tab_part_n11 partition_columns ds partition_columns.types string - serialization.ddl struct tab_part_n11 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tab_part_n11 name: default.tab_part_n11 @@ -1046,25 +988,17 @@ STAGE PLANS: partition values: ds 2008-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.tab_n10 - numFiles 2 - numRows 150 partition_columns ds partition_columns.types string - rawDataSize 1598 - serialization.ddl struct tab_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1748 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -1081,10 +1015,8 @@ STAGE PLANS: name default.tab_n10 partition_columns ds partition_columns.types string - serialization.ddl struct tab_n10 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tab_n10 name: default.tab_n10 diff --git a/ql/src/test/results/clientpositive/llap/murmur_hash_migration2.q.out b/ql/src/test/results/clientpositive/llap/murmur_hash_migration2.q.out index 29f50d8556..cc74705cc5 100644 --- a/ql/src/test/results/clientpositive/llap/murmur_hash_migration2.q.out +++ b/ql/src/test/results/clientpositive/llap/murmur_hash_migration2.q.out @@ -75,11 +75,9 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns - columns.comments columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -87,7 +85,6 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.NullRowsInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns @@ -95,7 +92,6 @@ STAGE PLANS: columns.types #### A masked pattern was here #### name _dummy_database._dummy_table - serialization.ddl struct _dummy_table { } serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe serde: org.apache.hadoop.hive.serde2.NullStructSerDe @@ -134,12 +130,10 @@ STAGE PLANS: name default.acid_ptn_bucket1 partition_columns ds partition_columns.types string - serialization.ddl struct acid_ptn_bucket1 { i32 a, i32 b} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde transactional true transactional_properties default -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.acid_ptn_bucket1 TotalFiles: 1 @@ -172,12 +166,10 @@ STAGE PLANS: name default.acid_ptn_bucket1 partition_columns ds partition_columns.types string - serialization.ddl struct acid_ptn_bucket1 { i32 a, i32 b} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde transactional true transactional_properties default -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.acid_ptn_bucket1 Write Type: INSERT diff --git a/ql/src/test/results/clientpositive/llap/offset_limit_global_optimizer.q.out b/ql/src/test/results/clientpositive/llap/offset_limit_global_optimizer.q.out index bf5c28670d..7e14fc7988 100644 --- a/ql/src/test/results/clientpositive/llap/offset_limit_global_optimizer.q.out +++ b/ql/src/test/results/clientpositive/llap/offset_limit_global_optimizer.q.out @@ -68,30 +68,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -101,10 +91,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -117,30 +105,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -150,10 +128,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -166,30 +142,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -199,10 +165,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -215,30 +179,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -248,10 +202,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -399,30 +351,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -432,10 +374,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -448,30 +388,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -481,10 +411,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -497,30 +425,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -530,10 +448,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -546,30 +462,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -579,10 +485,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -730,30 +634,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -763,10 +657,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -779,30 +671,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -812,10 +694,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -828,30 +708,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -861,10 +731,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -877,30 +745,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -910,10 +768,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1071,30 +927,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1104,10 +950,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1120,30 +964,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1153,10 +987,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1169,30 +1001,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1202,10 +1024,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1218,30 +1038,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1251,10 +1061,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1987,30 +1795,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2020,10 +1818,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2036,30 +1832,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2069,10 +1855,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2085,30 +1869,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2118,10 +1892,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2134,30 +1906,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2167,10 +1929,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2313,30 +2073,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2346,10 +2096,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2362,30 +2110,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2395,10 +2133,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2411,30 +2147,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2444,10 +2170,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2460,30 +2184,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2493,10 +2207,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2639,30 +2351,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2672,10 +2374,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2688,30 +2388,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2721,10 +2411,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2737,30 +2425,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2770,10 +2448,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2786,30 +2462,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2819,10 +2485,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -2975,30 +2639,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3008,10 +2662,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -3024,30 +2676,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3057,10 +2699,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -3073,30 +2713,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3106,10 +2736,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -3122,30 +2750,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3155,10 +2773,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/llap/optimize_nullscan.q.out b/ql/src/test/results/clientpositive/llap/optimize_nullscan.q.out index 77ddaa273e..b1a46e995a 100644 --- a/ql/src/test/results/clientpositive/llap/optimize_nullscan.q.out +++ b/ql/src/test/results/clientpositive/llap/optimize_nullscan.q.out @@ -206,30 +206,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -237,14 +227,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -289,30 +273,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -322,10 +296,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -337,30 +309,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -370,10 +332,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -385,30 +345,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -418,10 +368,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -433,30 +381,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -466,10 +404,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -608,30 +544,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -639,14 +565,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -691,30 +611,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -724,10 +634,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -740,30 +648,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -773,10 +671,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -789,30 +685,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -822,10 +708,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -838,30 +722,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -871,10 +745,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1044,30 +916,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1075,14 +937,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -1126,30 +982,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1159,10 +1005,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1174,30 +1018,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1207,10 +1041,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1222,30 +1054,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1255,10 +1077,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1270,30 +1090,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1303,10 +1113,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1452,30 +1260,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1483,14 +1281,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -1526,30 +1318,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1557,14 +1339,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -1701,30 +1477,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1732,14 +1498,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -1776,30 +1536,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1807,14 +1557,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -1933,30 +1677,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.OneNullRowInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.NullStructSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.NullStructSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1964,14 +1698,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/llap/outer_join_ppr.q.out b/ql/src/test/results/clientpositive/llap/outer_join_ppr.q.out index d7bc092156..13bfa0fb2d 100644 --- a/ql/src/test/results/clientpositive/llap/outer_join_ppr.q.out +++ b/ql/src/test/results/clientpositive/llap/outer_join_ppr.q.out @@ -82,30 +82,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -113,14 +103,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -166,30 +150,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -199,10 +173,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -215,30 +187,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -248,10 +210,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -424,30 +384,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -455,14 +405,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -508,30 +452,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -541,10 +475,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -557,30 +489,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -590,10 +512,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/llap/parquet_vectorization_0.q.out b/ql/src/test/results/clientpositive/llap/parquet_vectorization_0.q.out index 8b5e892d9a..93ecd40c36 100644 --- a/ql/src/test/results/clientpositive/llap/parquet_vectorization_0.q.out +++ b/ql/src/test/results/clientpositive/llap/parquet_vectorization_0.q.out @@ -1263,30 +1263,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat output format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 - columns.comments columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesparquet - numFiles 1 - numRows 12288 - rawDataSize __SOME_NUMBER__ - serialization.ddl struct alltypesparquet { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe - totalSize __SOME_NUMBER__ -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe input format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat output format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 @@ -1294,14 +1284,8 @@ STAGE PLANS: columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesparquet - numFiles 1 - numRows 12288 - rawDataSize __SOME_NUMBER__ - serialization.ddl struct alltypesparquet { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe - totalSize __SOME_NUMBER__ -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe name: default.alltypesparquet name: default.alltypesparquet @@ -30098,30 +30082,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat output format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 - columns.comments columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesparquet - numFiles 1 - numRows 12288 - rawDataSize __SOME_NUMBER__ - serialization.ddl struct alltypesparquet { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe - totalSize __SOME_NUMBER__ -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe input format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat output format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 @@ -30129,14 +30103,8 @@ STAGE PLANS: columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesparquet - numFiles 1 - numRows 12288 - rawDataSize __SOME_NUMBER__ - serialization.ddl struct alltypesparquet { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe - totalSize __SOME_NUMBER__ -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe name: default.alltypesparquet name: default.alltypesparquet @@ -30225,30 +30193,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat output format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 - columns.comments columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesparquet - numFiles 1 - numRows 12288 - rawDataSize __SOME_NUMBER__ - serialization.ddl struct alltypesparquet { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe - totalSize __SOME_NUMBER__ -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe input format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat output format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 @@ -30256,14 +30214,8 @@ STAGE PLANS: columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesparquet - numFiles 1 - numRows 12288 - rawDataSize __SOME_NUMBER__ - serialization.ddl struct alltypesparquet { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe - totalSize __SOME_NUMBER__ -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe name: default.alltypesparquet name: default.alltypesparquet @@ -30352,30 +30304,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat output format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 - columns.comments columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesparquet - numFiles 1 - numRows 12288 - rawDataSize __SOME_NUMBER__ - serialization.ddl struct alltypesparquet { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe - totalSize __SOME_NUMBER__ -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe input format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat output format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 @@ -30383,14 +30325,8 @@ STAGE PLANS: columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesparquet - numFiles 1 - numRows 12288 - rawDataSize __SOME_NUMBER__ - serialization.ddl struct alltypesparquet { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe - totalSize __SOME_NUMBER__ -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe name: default.alltypesparquet name: default.alltypesparquet @@ -30473,30 +30409,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat output format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 - columns.comments columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesparquet - numFiles 1 - numRows 12288 - rawDataSize __SOME_NUMBER__ - serialization.ddl struct alltypesparquet { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe - totalSize __SOME_NUMBER__ -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe input format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetInputFormat output format: org.apache.hadoop.hive.ql.io.parquet.MapredParquetOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 @@ -30504,14 +30430,8 @@ STAGE PLANS: columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesparquet - numFiles 1 - numRows 12288 - rawDataSize __SOME_NUMBER__ - serialization.ddl struct alltypesparquet { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe - totalSize __SOME_NUMBER__ -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe name: default.alltypesparquet name: default.alltypesparquet diff --git a/ql/src/test/results/clientpositive/llap/partition_pruning.q.out b/ql/src/test/results/clientpositive/llap/partition_pruning.q.out index 1382266774..362dda0abd 100644 --- a/ql/src/test/results/clientpositive/llap/partition_pruning.q.out +++ b/ql/src/test/results/clientpositive/llap/partition_pruning.q.out @@ -119,30 +119,20 @@ STAGE PLANS: partition values: dt 2001-01-01 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 column.name.delimiter , columns customer - columns.comments columns.types int #### A masked pattern was here #### name default.daysales - numFiles 1 - numRows 1 partition_columns dt partition_columns.types string - rawDataSize 1 - serialization.ddl struct daysales { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns customer @@ -152,10 +142,8 @@ STAGE PLANS: name default.daysales partition_columns dt partition_columns.types string - serialization.ddl struct daysales { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.daysales name: default.daysales @@ -201,30 +189,20 @@ STAGE PLANS: partition values: dt 2001-01-01 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 column.name.delimiter , columns customer - columns.comments columns.types int #### A masked pattern was here #### name default.daysales - numFiles 1 - numRows 1 partition_columns dt partition_columns.types string - rawDataSize 1 - serialization.ddl struct daysales { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns customer @@ -234,10 +212,8 @@ STAGE PLANS: name default.daysales partition_columns dt partition_columns.types string - serialization.ddl struct daysales { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.daysales name: default.daysales @@ -247,30 +223,20 @@ STAGE PLANS: partition values: dt 2001-01-03 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 column.name.delimiter , columns customer - columns.comments columns.types int #### A masked pattern was here #### name default.daysales - numFiles 1 - numRows 1 partition_columns dt partition_columns.types string - rawDataSize 1 - serialization.ddl struct daysales { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns customer @@ -280,10 +246,8 @@ STAGE PLANS: name default.daysales partition_columns dt partition_columns.types string - serialization.ddl struct daysales { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.daysales name: default.daysales @@ -329,30 +293,20 @@ STAGE PLANS: partition values: dt 2001-01-01 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 column.name.delimiter , columns customer - columns.comments columns.types int #### A masked pattern was here #### name default.daysales - numFiles 1 - numRows 1 partition_columns dt partition_columns.types string - rawDataSize 1 - serialization.ddl struct daysales { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns customer @@ -362,10 +316,8 @@ STAGE PLANS: name default.daysales partition_columns dt partition_columns.types string - serialization.ddl struct daysales { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.daysales name: default.daysales @@ -375,30 +327,20 @@ STAGE PLANS: partition values: dt 2001-01-03 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 column.name.delimiter , columns customer - columns.comments columns.types int #### A masked pattern was here #### name default.daysales - numFiles 1 - numRows 1 partition_columns dt partition_columns.types string - rawDataSize 1 - serialization.ddl struct daysales { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns customer @@ -408,10 +350,8 @@ STAGE PLANS: name default.daysales partition_columns dt partition_columns.types string - serialization.ddl struct daysales { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.daysales name: default.daysales diff --git a/ql/src/test/results/clientpositive/llap/pcr.q.out b/ql/src/test/results/clientpositive/llap/pcr.q.out index 755acde898..d10364c727 100644 --- a/ql/src/test/results/clientpositive/llap/pcr.q.out +++ b/ql/src/test/results/clientpositive/llap/pcr.q.out @@ -114,30 +114,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -147,10 +137,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -162,30 +150,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -195,10 +173,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -334,30 +310,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -367,10 +333,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -382,30 +346,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -415,10 +369,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -430,30 +382,20 @@ STAGE PLANS: partition values: ds 2000-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -463,10 +405,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -638,30 +578,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -671,10 +601,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -686,30 +614,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -719,10 +637,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -858,30 +774,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -891,10 +797,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -906,30 +810,20 @@ STAGE PLANS: partition values: ds 2000-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -939,10 +833,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -1082,30 +974,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1115,10 +997,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -1130,30 +1010,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1163,10 +1033,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -1178,30 +1046,20 @@ STAGE PLANS: partition values: ds 2000-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1211,10 +1069,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -1365,30 +1221,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1398,10 +1244,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -1413,30 +1257,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1446,10 +1280,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -1461,30 +1293,20 @@ STAGE PLANS: partition values: ds 2000-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1494,10 +1316,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -1654,30 +1474,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1687,10 +1497,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -1702,30 +1510,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1735,10 +1533,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -1857,30 +1653,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1890,10 +1676,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -1905,30 +1689,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1938,10 +1712,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -2102,30 +1874,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2135,10 +1897,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -2150,30 +1910,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2183,10 +1933,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -2198,30 +1946,20 @@ STAGE PLANS: partition values: ds 2000-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2231,10 +1969,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -2420,30 +2156,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2453,10 +2179,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -2468,30 +2192,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2501,10 +2215,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -2634,30 +2346,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2667,10 +2369,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -2715,30 +2415,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2748,10 +2438,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -2940,30 +2628,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2973,10 +2651,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -3021,30 +2697,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3054,10 +2720,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -3254,30 +2918,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3287,10 +2941,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -3302,30 +2954,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3335,10 +2977,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -3350,30 +2990,20 @@ STAGE PLANS: partition values: ds 2000-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3383,10 +3013,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -3398,30 +3026,20 @@ STAGE PLANS: partition values: ds 2000-04-11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3431,10 +3049,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -3607,30 +3223,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3640,10 +3246,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -3655,30 +3259,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3688,10 +3282,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -3703,30 +3295,20 @@ STAGE PLANS: partition values: ds 2000-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3736,10 +3318,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -3914,8 +3494,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3923,14 +3501,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.pcr_t2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct pcr_t2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2 TotalFiles: 1 @@ -3971,8 +3543,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -3980,14 +3550,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.pcr_t3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct pcr_t3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t3 TotalFiles: 1 @@ -4025,30 +3589,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4058,10 +3612,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -4146,8 +3698,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4155,14 +3705,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.pcr_t2 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct pcr_t2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2 @@ -4185,8 +3729,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4194,14 +3736,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.pcr_t3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct pcr_t3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t3 @@ -4296,8 +3832,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4305,14 +3839,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.pcr_t2 - numFiles 1 - numRows 20 - rawDataSize 160 - serialization.ddl struct pcr_t2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2 TotalFiles: 1 @@ -4357,8 +3885,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4366,14 +3892,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.pcr_t3 - numFiles 1 - numRows 20 - rawDataSize 160 - serialization.ddl struct pcr_t3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t3 TotalFiles: 1 @@ -4411,30 +3931,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4444,10 +3954,8 @@ STAGE PLANS: name default.pcr_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1 name: default.pcr_t1 @@ -4532,8 +4040,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4541,14 +4047,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.pcr_t2 - numFiles 1 - numRows 20 - rawDataSize 160 - serialization.ddl struct pcr_t2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2 @@ -4571,8 +4071,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4580,14 +4078,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.pcr_t3 - numFiles 1 - numRows 20 - rawDataSize 160 - serialization.ddl struct pcr_t3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t3 @@ -4691,30 +4183,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4724,10 +4206,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -4851,30 +4331,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4884,10 +4354,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -4900,30 +4368,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -4933,10 +4391,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -5064,30 +4520,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -5097,10 +4543,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -5113,30 +4557,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -5146,10 +4580,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/llap/pcs.q.out b/ql/src/test/results/clientpositive/llap/pcs.q.out index 9a1d5a5362..c20618e28e 100644 --- a/ql/src/test/results/clientpositive/llap/pcs.q.out +++ b/ql/src/test/results/clientpositive/llap/pcs.q.out @@ -151,30 +151,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -184,10 +174,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -199,30 +187,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -232,10 +210,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -323,30 +299,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -356,10 +322,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -369,30 +333,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -402,10 +356,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -464,30 +416,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -497,10 +439,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -510,30 +450,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -543,10 +473,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -655,30 +583,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -688,10 +606,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -703,30 +619,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -736,10 +642,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -781,30 +685,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -814,10 +708,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -829,30 +721,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -862,10 +744,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -962,30 +842,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -995,10 +865,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -1008,30 +876,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1041,10 +899,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -1353,30 +1209,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1386,10 +1232,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -1462,30 +1306,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1495,10 +1329,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -1508,30 +1340,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1541,10 +1363,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -1591,30 +1411,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1624,10 +1434,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -1637,30 +1445,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1670,10 +1468,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -1683,30 +1479,20 @@ STAGE PLANS: partition values: ds 2000-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1716,10 +1502,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -1784,30 +1568,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1817,10 +1591,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -1830,30 +1602,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1863,10 +1625,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 @@ -1876,30 +1636,20 @@ STAGE PLANS: partition values: ds 2000-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcs_t1 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1909,10 +1659,8 @@ STAGE PLANS: name default.pcs_t1 partition_columns ds partition_columns.types string - serialization.ddl struct pcs_t1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcs_t1 name: default.pcs_t1 diff --git a/ql/src/test/results/clientpositive/llap/pointlookup2.q.out b/ql/src/test/results/clientpositive/llap/pointlookup2.q.out index 01fadb3c62..ef413831e3 100644 --- a/ql/src/test/results/clientpositive/llap/pointlookup2.q.out +++ b/ql/src/test/results/clientpositive/llap/pointlookup2.q.out @@ -159,30 +159,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -192,10 +182,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -207,30 +195,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -240,10 +218,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -368,30 +344,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -401,10 +367,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -449,30 +413,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -482,10 +436,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -634,30 +586,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -667,10 +609,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -715,30 +655,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -748,10 +678,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -896,30 +824,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -929,10 +847,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -944,30 +860,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -977,10 +883,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -1022,30 +926,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ds,key,value - columns.comments columns.types string:int:string #### A masked pattern was here #### name default.pcr_t2_n0 - numFiles 1 - numRows 1 - rawDataSize 18 - serialization.ddl struct pcr_t2_n0 { string ds, i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 19 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ds,key,value @@ -1053,14 +947,8 @@ STAGE PLANS: columns.types string:int:string #### A masked pattern was here #### name default.pcr_t2_n0 - numFiles 1 - numRows 1 - rawDataSize 18 - serialization.ddl struct pcr_t2_n0 { string ds, i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 19 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2_n0 name: default.pcr_t2_n0 @@ -1216,30 +1104,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1249,10 +1127,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -1264,30 +1140,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1297,10 +1163,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -1312,30 +1176,20 @@ STAGE PLANS: partition values: ds 2000-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1345,10 +1199,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -1391,30 +1243,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ds,key,value - columns.comments columns.types string:int:string #### A masked pattern was here #### name default.pcr_t2_n0 - numFiles 1 - numRows 1 - rawDataSize 18 - serialization.ddl struct pcr_t2_n0 { string ds, i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 19 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ds,key,value @@ -1422,14 +1264,8 @@ STAGE PLANS: columns.types string:int:string #### A masked pattern was here #### name default.pcr_t2_n0 - numFiles 1 - numRows 1 - rawDataSize 18 - serialization.ddl struct pcr_t2_n0 { string ds, i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 19 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2_n0 name: default.pcr_t2_n0 @@ -1753,30 +1589,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1786,10 +1612,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -1801,30 +1625,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1834,10 +1648,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -1962,30 +1774,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1995,10 +1797,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -2043,30 +1843,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2076,10 +1866,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -2228,30 +2016,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2261,10 +2039,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -2309,30 +2085,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2342,10 +2108,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -2482,30 +2246,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2515,10 +2269,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -2530,30 +2282,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2563,10 +2305,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -2608,30 +2348,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ds,key,value - columns.comments columns.types string:int:string #### A masked pattern was here #### name default.pcr_t2_n0 - numFiles 1 - numRows 1 - rawDataSize 18 - serialization.ddl struct pcr_t2_n0 { string ds, i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 19 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ds,key,value @@ -2639,14 +2369,8 @@ STAGE PLANS: columns.types string:int:string #### A masked pattern was here #### name default.pcr_t2_n0 - numFiles 1 - numRows 1 - rawDataSize 18 - serialization.ddl struct pcr_t2_n0 { string ds, i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 19 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2_n0 name: default.pcr_t2_n0 @@ -2790,30 +2514,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2823,10 +2537,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -2838,30 +2550,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2871,10 +2573,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -2886,30 +2586,20 @@ STAGE PLANS: partition values: ds 2000-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n2 - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2919,10 +2609,8 @@ STAGE PLANS: name default.pcr_t1_n2 partition_columns ds partition_columns.types string - serialization.ddl struct pcr_t1_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n2 name: default.pcr_t1_n2 @@ -2965,30 +2653,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ds,key,value - columns.comments columns.types string:int:string #### A masked pattern was here #### name default.pcr_t2_n0 - numFiles 1 - numRows 1 - rawDataSize 18 - serialization.ddl struct pcr_t2_n0 { string ds, i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 19 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ds,key,value @@ -2996,14 +2674,8 @@ STAGE PLANS: columns.types string:int:string #### A masked pattern was here #### name default.pcr_t2_n0 - numFiles 1 - numRows 1 - rawDataSize 18 - serialization.ddl struct pcr_t2_n0 { string ds, i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 19 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t2_n0 name: default.pcr_t2_n0 diff --git a/ql/src/test/results/clientpositive/llap/pointlookup3.q.out b/ql/src/test/results/clientpositive/llap/pointlookup3.q.out index d945be2023..7962d3391c 100644 --- a/ql/src/test/results/clientpositive/llap/pointlookup3.q.out +++ b/ql/src/test/results/clientpositive/llap/pointlookup3.q.out @@ -114,30 +114,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -147,10 +137,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -163,30 +151,20 @@ STAGE PLANS: ds1 2000-04-09 ds2 2001-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -196,10 +174,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -318,30 +294,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -351,10 +317,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -479,30 +443,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -512,10 +466,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -561,30 +513,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -594,10 +536,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -747,30 +687,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -780,10 +710,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -829,30 +757,20 @@ STAGE PLANS: ds1 2000-04-09 ds2 2001-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -862,10 +780,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -1011,30 +927,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1044,10 +950,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -1060,30 +964,20 @@ STAGE PLANS: ds1 2000-04-09 ds2 2001-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1093,10 +987,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -1141,30 +1033,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1174,10 +1056,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -1190,30 +1070,20 @@ STAGE PLANS: ds1 2000-04-09 ds2 2001-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1223,10 +1093,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -1239,30 +1107,20 @@ STAGE PLANS: ds1 2000-04-10 ds2 2001-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1272,10 +1130,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -1650,30 +1506,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1683,10 +1529,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -1699,30 +1543,20 @@ STAGE PLANS: ds1 2000-04-09 ds2 2001-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1732,10 +1566,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -1854,30 +1686,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1887,10 +1709,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -2015,30 +1835,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2048,10 +1858,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -2097,30 +1905,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2130,10 +1928,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -2283,30 +2079,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2316,10 +2102,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -2365,30 +2149,20 @@ STAGE PLANS: ds1 2000-04-09 ds2 2001-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2398,10 +2172,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -2539,30 +2311,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2572,10 +2334,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -2588,30 +2348,20 @@ STAGE PLANS: ds1 2000-04-09 ds2 2001-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2621,10 +2371,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -2669,30 +2417,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2702,10 +2440,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -2718,30 +2454,20 @@ STAGE PLANS: ds1 2000-04-09 ds2 2001-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2751,10 +2477,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 @@ -2767,30 +2491,20 @@ STAGE PLANS: ds1 2000-04-10 ds2 2001-04-10 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n1 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -2800,10 +2514,8 @@ STAGE PLANS: name default.pcr_t1_n1 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n1 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n1 name: default.pcr_t1_n1 diff --git a/ql/src/test/results/clientpositive/llap/pointlookup4.q.out b/ql/src/test/results/clientpositive/llap/pointlookup4.q.out index 3ca21d7460..0eb6a5be48 100644 --- a/ql/src/test/results/clientpositive/llap/pointlookup4.q.out +++ b/ql/src/test/results/clientpositive/llap/pointlookup4.q.out @@ -114,30 +114,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n0 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -147,10 +137,8 @@ STAGE PLANS: name default.pcr_t1_n0 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n0 name: default.pcr_t1_n0 @@ -163,30 +151,20 @@ STAGE PLANS: ds1 2000-04-09 ds2 2001-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n0 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -196,10 +174,8 @@ STAGE PLANS: name default.pcr_t1_n0 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n0 name: default.pcr_t1_n0 @@ -338,30 +314,20 @@ STAGE PLANS: ds1 2000-04-08 ds2 2001-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n0 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -371,10 +337,8 @@ STAGE PLANS: name default.pcr_t1_n0 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n0 name: default.pcr_t1_n0 @@ -387,30 +351,20 @@ STAGE PLANS: ds1 2000-04-09 ds2 2001-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.pcr_t1_n0 - numFiles 1 - numRows 20 partition_columns ds1/ds2 partition_columns.types string:string - rawDataSize 160 - serialization.ddl struct pcr_t1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -420,10 +374,8 @@ STAGE PLANS: name default.pcr_t1_n0 partition_columns ds1/ds2 partition_columns.types string:string - serialization.ddl struct pcr_t1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.pcr_t1_n0 name: default.pcr_t1_n0 diff --git a/ql/src/test/results/clientpositive/llap/ppd_join_filter.q.out b/ql/src/test/results/clientpositive/llap/ppd_join_filter.q.out index 5145494c27..52db4aaa1c 100644 --- a/ql/src/test/results/clientpositive/llap/ppd_join_filter.q.out +++ b/ql/src/test/results/clientpositive/llap/ppd_join_filter.q.out @@ -104,30 +104,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -135,14 +125,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -365,30 +349,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -396,14 +370,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -625,30 +593,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -656,14 +614,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -886,30 +838,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -917,14 +859,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/llap/ppd_union_view.q.out b/ql/src/test/results/clientpositive/llap/ppd_union_view.q.out index d16d28b64b..036474b446 100644 --- a/ql/src/test/results/clientpositive/llap/ppd_union_view.q.out +++ b/ql/src/test/results/clientpositive/llap/ppd_union_view.q.out @@ -261,30 +261,20 @@ STAGE PLANS: partition values: ds 2011-10-13 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"keymap":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns keymap,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_old - numFiles 1 - numRows 1 partition_columns ds partition_columns.types string - rawDataSize 14 - serialization.ddl struct t1_old { string keymap, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 15 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns keymap,value @@ -296,10 +286,8 @@ STAGE PLANS: name default.t1_old partition_columns ds partition_columns.types string - serialization.ddl struct t1_old { string keymap, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_old name: default.t1_old @@ -346,30 +334,20 @@ STAGE PLANS: partition values: ds 2011-10-13 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","keymap":"true"}} - bucket_count -1 column.name.delimiter , columns key,keymap - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_mapping - numFiles 1 - numRows 1 partition_columns ds partition_columns.types string - rawDataSize 12 - serialization.ddl struct t1_mapping { string key, string keymap} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 13 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,keymap @@ -381,10 +359,8 @@ STAGE PLANS: name default.t1_mapping partition_columns ds partition_columns.types string - serialization.ddl struct t1_mapping { string key, string keymap} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_mapping name: default.t1_mapping @@ -577,30 +553,20 @@ STAGE PLANS: partition values: ds 2011-10-15 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.t1_new_n0 - numFiles 1 - numRows 1 partition_columns ds partition_columns.types string - rawDataSize 11 - serialization.ddl struct t1_new_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 12 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -612,10 +578,8 @@ STAGE PLANS: name default.t1_new_n0 partition_columns ds partition_columns.types string - serialization.ddl struct t1_new_n0 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.t1_new_n0 name: default.t1_new_n0 diff --git a/ql/src/test/results/clientpositive/llap/ppd_vc.q.out b/ql/src/test/results/clientpositive/llap/ppd_vc.q.out index ebb3363172..8df5077aff 100644 --- a/ql/src/test/results/clientpositive/llap/ppd_vc.q.out +++ b/ql/src/test/results/clientpositive/llap/ppd_vc.q.out @@ -34,30 +34,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -67,10 +57,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -81,30 +69,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -114,10 +92,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -128,30 +104,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -161,10 +127,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -175,30 +139,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -208,10 +162,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -361,30 +313,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -392,14 +334,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -445,30 +381,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -478,10 +404,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -494,30 +418,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -527,10 +441,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -543,30 +455,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -576,10 +478,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -592,30 +492,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -625,10 +515,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/llap/ppr_allchildsarenull.q.out b/ql/src/test/results/clientpositive/llap/ppr_allchildsarenull.q.out index 0061e60fd4..ebfedae2fe 100644 --- a/ql/src/test/results/clientpositive/llap/ppr_allchildsarenull.q.out +++ b/ql/src/test/results/clientpositive/llap/ppr_allchildsarenull.q.out @@ -42,30 +42,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -75,10 +65,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -89,30 +77,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -122,10 +100,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -214,30 +190,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -247,10 +213,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -261,30 +225,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -294,10 +248,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -308,30 +260,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -341,10 +283,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -355,30 +295,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -388,10 +318,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/llap/push_or.q.out b/ql/src/test/results/clientpositive/llap/push_or.q.out index 1ac850df8f..9df0ab7239 100644 --- a/ql/src/test/results/clientpositive/llap/push_or.q.out +++ b/ql/src/test/results/clientpositive/llap/push_or.q.out @@ -96,30 +96,20 @@ STAGE PLANS: partition values: ds 2000-04-08 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.push_or - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct push_or { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -129,10 +119,8 @@ STAGE PLANS: name default.push_or partition_columns ds partition_columns.types string - serialization.ddl struct push_or { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.push_or name: default.push_or @@ -144,30 +132,20 @@ STAGE PLANS: partition values: ds 2000-04-09 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.push_or - numFiles 1 - numRows 20 partition_columns ds partition_columns.types string - rawDataSize 160 - serialization.ddl struct push_or { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 180 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -177,10 +155,8 @@ STAGE PLANS: name default.push_or partition_columns ds partition_columns.types string - serialization.ddl struct push_or { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.push_or name: default.push_or diff --git a/ql/src/test/results/clientpositive/llap/rand_partitionpruner2.q.out b/ql/src/test/results/clientpositive/llap/rand_partitionpruner2.q.out index ef5509281a..13249e91f8 100644 --- a/ql/src/test/results/clientpositive/llap/rand_partitionpruner2.q.out +++ b/ql/src/test/results/clientpositive/llap/rand_partitionpruner2.q.out @@ -65,8 +65,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,hr,ds @@ -74,14 +72,8 @@ STAGE PLANS: columns.types string:string:string:string #### A masked pattern was here #### name default.tmptable_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct tmptable_n1 { string key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tmptable_n1 TotalFiles: 1 @@ -120,30 +112,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -153,10 +135,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -169,30 +149,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -202,10 +172,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -258,8 +226,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"ds":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,hr,ds @@ -267,14 +233,8 @@ STAGE PLANS: columns.types string:string:string:string #### A masked pattern was here #### name default.tmptable_n1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct tmptable_n1 { string key, string value, string hr, string ds} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.tmptable_n1 diff --git a/ql/src/test/results/clientpositive/llap/rand_partitionpruner3.q.out b/ql/src/test/results/clientpositive/llap/rand_partitionpruner3.q.out index b7f387c05c..7f3aae3507 100644 --- a/ql/src/test/results/clientpositive/llap/rand_partitionpruner3.q.out +++ b/ql/src/test/results/clientpositive/llap/rand_partitionpruner3.q.out @@ -26,30 +26,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -59,10 +49,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -118,30 +106,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -151,10 +129,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/llap/reduce_deduplicate.q.out b/ql/src/test/results/clientpositive/llap/reduce_deduplicate.q.out index 29c233fcda..33ebf0cb14 100644 --- a/ql/src/test/results/clientpositive/llap/reduce_deduplicate.q.out +++ b/ql/src/test/results/clientpositive/llap/reduce_deduplicate.q.out @@ -64,31 +64,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -97,14 +87,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -130,7 +114,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -141,14 +124,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.bucket5_1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket5_1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket5_1 TotalFiles: 2 @@ -200,7 +177,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -211,14 +187,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.bucket5_1 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucket5_1 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucket5_1 @@ -403,7 +373,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns aid,bid,t,ctime,etime,l,et @@ -414,10 +383,8 @@ STAGE PLANS: name default.complex_tbl_1 partition_columns ds partition_columns.types string - serialization.ddl struct complex_tbl_1 { string aid, string bid, i32 t, string ctime, i64 etime, string l, string et} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.complex_tbl_1 TotalFiles: 1 @@ -498,7 +465,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns aid,bid,t,ctime,etime,l,et @@ -509,10 +475,8 @@ STAGE PLANS: name default.complex_tbl_1 partition_columns ds partition_columns.types string - serialization.ddl struct complex_tbl_1 { string aid, string bid, i32 t, string ctime, i64 etime, string l, string et} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.complex_tbl_1 diff --git a/ql/src/test/results/clientpositive/llap/router_join_ppr.q.out b/ql/src/test/results/clientpositive/llap/router_join_ppr.q.out index de20bb6209..8180eb7759 100644 --- a/ql/src/test/results/clientpositive/llap/router_join_ppr.q.out +++ b/ql/src/test/results/clientpositive/llap/router_join_ppr.q.out @@ -82,30 +82,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -113,14 +103,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -166,30 +150,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -199,10 +173,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -215,30 +187,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -248,10 +210,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -427,30 +387,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -460,10 +410,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -476,30 +424,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -509,10 +447,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -556,30 +492,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -587,14 +513,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -766,30 +686,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -797,14 +707,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -850,30 +754,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -883,10 +777,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -899,30 +791,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -932,10 +814,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1111,30 +991,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1144,10 +1014,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1160,30 +1028,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1193,10 +1051,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -1240,30 +1096,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1271,14 +1117,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/llap/sample1.q.out b/ql/src/test/results/clientpositive/llap/sample1.q.out index 81a821d906..b257edc45c 100644 --- a/ql/src/test/results/clientpositive/llap/sample1.q.out +++ b/ql/src/test/results/clientpositive/llap/sample1.q.out @@ -63,8 +63,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,dt,hr @@ -72,14 +70,8 @@ STAGE PLANS: columns.types int:string:string:string #### A masked pattern was here #### name default.dest1_n89 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n89 { i32 key, string value, string dt, string hr} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n89 TotalFiles: 1 @@ -118,30 +110,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -151,10 +133,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -206,8 +186,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"dt":"true","hr":"true","key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value,dt,hr @@ -215,14 +193,8 @@ STAGE PLANS: columns.types int:string:string:string #### A masked pattern was here #### name default.dest1_n89 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n89 { i32 key, string value, string dt, string hr} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n89 diff --git a/ql/src/test/results/clientpositive/llap/sample10.q.out b/ql/src/test/results/clientpositive/llap/sample10.q.out index e1226296c9..5cb190cfc2 100644 --- a/ql/src/test/results/clientpositive/llap/sample10.q.out +++ b/ql/src/test/results/clientpositive/llap/sample10.q.out @@ -112,25 +112,17 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.srcpartbucket - numFiles 3 - numRows 10 partition_columns ds/hr partition_columns.types string:string - rawDataSize 60 - serialization.ddl struct srcpartbucket { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 295 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat @@ -147,10 +139,8 @@ STAGE PLANS: name default.srcpartbucket partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpartbucket { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.srcpartbucket name: default.srcpartbucket @@ -163,25 +153,17 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.srcpartbucket - numFiles 3 - numRows 10 partition_columns ds/hr partition_columns.types string:string - rawDataSize 60 - serialization.ddl struct srcpartbucket { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 295 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat @@ -198,10 +180,8 @@ STAGE PLANS: name default.srcpartbucket partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpartbucket { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.srcpartbucket name: default.srcpartbucket @@ -214,25 +194,17 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.srcpartbucket - numFiles 3 - numRows 10 partition_columns ds/hr partition_columns.types string:string - rawDataSize 60 - serialization.ddl struct srcpartbucket { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 295 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat @@ -249,10 +221,8 @@ STAGE PLANS: name default.srcpartbucket partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpartbucket { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.srcpartbucket name: default.srcpartbucket @@ -265,25 +235,17 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 4 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.srcpartbucket - numFiles 3 - numRows 10 partition_columns ds/hr partition_columns.types string:string - rawDataSize 60 - serialization.ddl struct srcpartbucket { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 295 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat @@ -300,10 +262,8 @@ STAGE PLANS: name default.srcpartbucket partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpartbucket { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.srcpartbucket name: default.srcpartbucket diff --git a/ql/src/test/results/clientpositive/llap/sample5.q.out b/ql/src/test/results/clientpositive/llap/sample5.q.out index d36a43679f..e870e8c9b2 100644 --- a/ql/src/test/results/clientpositive/llap/sample5.q.out +++ b/ql/src/test/results/clientpositive/llap/sample5.q.out @@ -60,8 +60,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -69,14 +67,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.dest1_n69 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n69 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n69 TotalFiles: 1 @@ -112,30 +104,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -145,14 +128,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket name: default.srcbucket @@ -204,8 +181,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -213,14 +188,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.dest1_n69 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n69 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n69 diff --git a/ql/src/test/results/clientpositive/llap/sample6.q.out b/ql/src/test/results/clientpositive/llap/sample6.q.out index cb4756329d..71141024c2 100644 --- a/ql/src/test/results/clientpositive/llap/sample6.q.out +++ b/ql/src/test/results/clientpositive/llap/sample6.q.out @@ -59,8 +59,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -68,14 +66,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.dest1_n27 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n27 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n27 TotalFiles: 1 @@ -111,30 +103,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -144,14 +127,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket name: default.srcbucket @@ -203,8 +180,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -212,14 +187,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.dest1_n27 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n27 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n27 @@ -581,30 +550,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -614,14 +574,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket name: default.srcbucket @@ -1005,30 +959,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -1038,14 +983,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket name: default.srcbucket @@ -1652,30 +1591,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -1685,14 +1615,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket name: default.srcbucket @@ -2181,30 +2105,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -2214,14 +2129,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket name: default.srcbucket @@ -2650,30 +2559,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 4 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket2 - numFiles 4 - numRows 500 - rawDataSize 5312 - serialization.ddl struct srcbucket2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 4 bucket_field_name key bucketing_version 2 @@ -2683,14 +2583,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket2 - numFiles 4 - numRows 500 - rawDataSize 5312 - serialization.ddl struct srcbucket2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket2 name: default.srcbucket2 @@ -3045,30 +2939,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 4 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket2 - numFiles 4 - numRows 500 - rawDataSize 5312 - serialization.ddl struct srcbucket2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 4 bucket_field_name key bucketing_version 2 @@ -3078,14 +2963,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket2 - numFiles 4 - numRows 500 - rawDataSize 5312 - serialization.ddl struct srcbucket2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket2 name: default.srcbucket2 @@ -3331,30 +3210,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.empty_bucket - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct empty_bucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -3364,14 +3234,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.empty_bucket - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct empty_bucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.empty_bucket name: default.empty_bucket diff --git a/ql/src/test/results/clientpositive/llap/sample7.q.out b/ql/src/test/results/clientpositive/llap/sample7.q.out index 369a4c6ef4..e217faa782 100644 --- a/ql/src/test/results/clientpositive/llap/sample7.q.out +++ b/ql/src/test/results/clientpositive/llap/sample7.q.out @@ -61,8 +61,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -70,14 +68,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.dest1_n160 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n160 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n160 TotalFiles: 1 @@ -113,30 +105,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key bucketing_version 2 @@ -146,14 +129,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket - numFiles 2 - numRows 1000 - rawDataSize 10603 - serialization.ddl struct srcbucket { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11603 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket name: default.srcbucket @@ -205,8 +182,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -214,14 +189,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.dest1_n160 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct dest1_n160 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dest1_n160 diff --git a/ql/src/test/results/clientpositive/llap/sample8.q.out b/ql/src/test/results/clientpositive/llap/sample8.q.out index cda918e8c4..3a1d2376bd 100644 --- a/ql/src/test/results/clientpositive/llap/sample8.q.out +++ b/ql/src/test/results/clientpositive/llap/sample8.q.out @@ -68,30 +68,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -101,10 +91,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -144,30 +132,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -177,10 +155,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -193,30 +169,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -226,10 +192,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -242,30 +206,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -275,10 +229,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -291,30 +243,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -324,10 +266,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/llap/sharedwork.q.out b/ql/src/test/results/clientpositive/llap/sharedwork.q.out index 175141fb9e..22b1cee6ac 100644 --- a/ql/src/test/results/clientpositive/llap/sharedwork.q.out +++ b/ql/src/test/results/clientpositive/llap/sharedwork.q.out @@ -168,30 +168,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col_1":"true","col_20":"true","col_3":"true","col_7":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns col_1,col_3,col_7,col_20 - columns.comments columns.types string:timestamp:string:string #### A masked pattern was here #### name default.my_table_0001 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct my_table_0001 { string col_1, timestamp col_3, string col_7, string col_20} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col_1":"true","col_20":"true","col_3":"true","col_7":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns col_1,col_3,col_7,col_20 @@ -199,14 +189,8 @@ STAGE PLANS: columns.types string:timestamp:string:string #### A masked pattern was here #### name default.my_table_0001 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct my_table_0001 { string col_1, timestamp col_3, string col_7, string col_20} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.my_table_0001 name: default.my_table_0001 @@ -268,30 +252,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col_21":"true","col_24":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns col_24,col_21 - columns.comments columns.types string:string #### A masked pattern was here #### name default.my_table_0003 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct my_table_0003 { string col_24, string col_21} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col_21":"true","col_24":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns col_24,col_21 @@ -299,14 +273,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.my_table_0003 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct my_table_0003 { string col_24, string col_21} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.my_table_0003 name: default.my_table_0003 @@ -349,30 +317,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col_1":"true","col_22":"true","col_23":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns col_1,col_22,col_23 - columns.comments columns.types string:string:int #### A masked pattern was here #### name default.my_table_0001_00 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct my_table_0001_00 { string col_1, string col_22, i32 col_23} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col_1":"true","col_22":"true","col_23":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns col_1,col_22,col_23 @@ -380,14 +338,8 @@ STAGE PLANS: columns.types string:string:int #### A masked pattern was here #### name default.my_table_0001_00 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct my_table_0001_00 { string col_1, string col_22, i32 col_23} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.my_table_0001_00 name: default.my_table_0001_00 @@ -429,30 +381,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col_1":"true","col_100":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns col_1,col_100 - columns.comments columns.types string:string #### A masked pattern was here #### name default.my_table_0001_01 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct my_table_0001_01 { string col_1, string col_100} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col_1":"true","col_100":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns col_1,col_100 @@ -460,14 +402,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.my_table_0001_01 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct my_table_0001_01 { string col_1, string col_100} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.my_table_0001_01 name: default.my_table_0001_01 @@ -722,30 +658,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"p_brand":"true","p_comment":"true","p_container":"true","p_mfgr":"true","p_name":"true","p_partkey":"true","p_retailprice":"true","p_size":"true","p_type":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns p_partkey,p_name,p_mfgr,p_brand,p_type,p_size,p_container,p_retailprice,p_comment - columns.comments columns.types int:string:string:string:string:int:string:double:string #### A masked pattern was here #### name default.part - numFiles 1 - numRows 26 - rawDataSize 3147 - serialization.ddl struct part { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3173 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"p_brand":"true","p_comment":"true","p_container":"true","p_mfgr":"true","p_name":"true","p_partkey":"true","p_retailprice":"true","p_size":"true","p_type":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns p_partkey,p_name,p_mfgr,p_brand,p_type,p_size,p_container,p_retailprice,p_comment @@ -753,14 +679,8 @@ STAGE PLANS: columns.types int:string:string:string:string:int:string:double:string #### A masked pattern was here #### name default.part - numFiles 1 - numRows 26 - rawDataSize 3147 - serialization.ddl struct part { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3173 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.part name: default.part @@ -834,30 +754,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"p_brand":"true","p_comment":"true","p_container":"true","p_mfgr":"true","p_name":"true","p_partkey":"true","p_retailprice":"true","p_size":"true","p_type":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns p_partkey,p_name,p_mfgr,p_brand,p_type,p_size,p_container,p_retailprice,p_comment - columns.comments columns.types int:string:string:string:string:int:string:double:string #### A masked pattern was here #### name default.part - numFiles 1 - numRows 26 - rawDataSize 3147 - serialization.ddl struct part { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3173 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"p_brand":"true","p_comment":"true","p_container":"true","p_mfgr":"true","p_name":"true","p_partkey":"true","p_retailprice":"true","p_size":"true","p_type":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns p_partkey,p_name,p_mfgr,p_brand,p_type,p_size,p_container,p_retailprice,p_comment @@ -865,14 +775,8 @@ STAGE PLANS: columns.types int:string:string:string:string:int:string:double:string #### A masked pattern was here #### name default.part - numFiles 1 - numRows 26 - rawDataSize 3147 - serialization.ddl struct part { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3173 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.part name: default.part @@ -930,30 +834,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"p_brand":"true","p_comment":"true","p_container":"true","p_mfgr":"true","p_name":"true","p_partkey":"true","p_retailprice":"true","p_size":"true","p_type":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns p_partkey,p_name,p_mfgr,p_brand,p_type,p_size,p_container,p_retailprice,p_comment - columns.comments columns.types int:string:string:string:string:int:string:double:string #### A masked pattern was here #### name default.part - numFiles 1 - numRows 26 - rawDataSize 3147 - serialization.ddl struct part { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3173 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"p_brand":"true","p_comment":"true","p_container":"true","p_mfgr":"true","p_name":"true","p_partkey":"true","p_retailprice":"true","p_size":"true","p_type":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns p_partkey,p_name,p_mfgr,p_brand,p_type,p_size,p_container,p_retailprice,p_comment @@ -961,14 +855,8 @@ STAGE PLANS: columns.types int:string:string:string:string:int:string:double:string #### A masked pattern was here #### name default.part - numFiles 1 - numRows 26 - rawDataSize 3147 - serialization.ddl struct part { i32 p_partkey, string p_name, string p_mfgr, string p_brand, string p_type, i32 p_size, string p_container, double p_retailprice, string p_comment} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 3173 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.part name: default.part diff --git a/ql/src/test/results/clientpositive/llap/smb_mapjoin_15.q.out b/ql/src/test/results/clientpositive/llap/smb_mapjoin_15.q.out index dbc180ccae..c2299c7503 100644 --- a/ql/src/test/results/clientpositive/llap/smb_mapjoin_15.q.out +++ b/ql/src/test/results/clientpositive/llap/smb_mapjoin_15.q.out @@ -89,31 +89,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.test_table2_n4 - numFiles 16 - numRows 500 - rawDataSize 5312 - serialization.ddl struct test_table2_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key @@ -124,14 +115,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.test_table2_n4 - numFiles 16 - numRows 500 - rawDataSize 5312 - serialization.ddl struct test_table2_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table2_n4 name: default.test_table2_n4 @@ -187,31 +172,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.test_table1_n4 - numFiles 16 - numRows 500 - rawDataSize 5312 - serialization.ddl struct test_table1_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key @@ -222,14 +198,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.test_table1_n4 - numFiles 16 - numRows 500 - rawDataSize 5312 - serialization.ddl struct test_table1_n4 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table1_n4 name: default.test_table1_n4 @@ -420,31 +390,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","key2":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,key2,value - columns.comments columns.types int:int:string #### A masked pattern was here #### name default.test_table1_n4 - numFiles 16 - numRows 500 - rawDataSize 7218 - serialization.ddl struct test_table1_n4 { i32 key, i32 key2, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7718 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","key2":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key @@ -455,14 +416,8 @@ STAGE PLANS: columns.types int:int:string #### A masked pattern was here #### name default.test_table1_n4 - numFiles 16 - numRows 500 - rawDataSize 7218 - serialization.ddl struct test_table1_n4 { i32 key, i32 key2, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7718 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table1_n4 name: default.test_table1_n4 @@ -505,31 +460,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","key2":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,key2,value - columns.comments columns.types int:int:string #### A masked pattern was here #### name default.test_table2_n4 - numFiles 16 - numRows 500 - rawDataSize 7218 - serialization.ddl struct test_table2_n4 { i32 key, i32 key2, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7718 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","key2":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key @@ -540,14 +486,8 @@ STAGE PLANS: columns.types int:int:string #### A masked pattern was here #### name default.test_table2_n4 - numFiles 16 - numRows 500 - rawDataSize 7218 - serialization.ddl struct test_table2_n4 { i32 key, i32 key2, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7718 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table2_n4 name: default.test_table2_n4 @@ -717,31 +657,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","key2":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,key2,value - columns.comments columns.types int:int:string #### A masked pattern was here #### name default.test_table1_n4 - numFiles 16 - numRows 500 - rawDataSize 7218 - serialization.ddl struct test_table1_n4 { i32 key, i32 key2, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7718 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","key2":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key @@ -752,14 +683,8 @@ STAGE PLANS: columns.types int:int:string #### A masked pattern was here #### name default.test_table1_n4 - numFiles 16 - numRows 500 - rawDataSize 7218 - serialization.ddl struct test_table1_n4 { i32 key, i32 key2, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7718 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table1_n4 name: default.test_table1_n4 @@ -802,31 +727,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","key2":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,key2,value - columns.comments columns.types int:int:string #### A masked pattern was here #### name default.test_table2_n4 - numFiles 16 - numRows 500 - rawDataSize 7218 - serialization.ddl struct test_table2_n4 { i32 key, i32 key2, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7718 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","key2":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key @@ -837,14 +753,8 @@ STAGE PLANS: columns.types int:int:string #### A masked pattern was here #### name default.test_table2_n4 - numFiles 16 - numRows 500 - rawDataSize 7218 - serialization.ddl struct test_table2_n4 { i32 key, i32 key2, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7718 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table2_n4 name: default.test_table2_n4 @@ -1014,31 +924,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","key2":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,key2,value - columns.comments columns.types int:int:string #### A masked pattern was here #### name default.test_table1_n4 - numFiles 16 - numRows 500 - rawDataSize 7218 - serialization.ddl struct test_table1_n4 { i32 key, i32 key2, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7718 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","key2":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key @@ -1049,14 +950,8 @@ STAGE PLANS: columns.types int:int:string #### A masked pattern was here #### name default.test_table1_n4 - numFiles 16 - numRows 500 - rawDataSize 7218 - serialization.ddl struct test_table1_n4 { i32 key, i32 key2, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7718 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table1_n4 name: default.test_table1_n4 @@ -1099,31 +994,22 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","key2":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key bucketing_version 2 column.name.delimiter , columns key,key2,value - columns.comments columns.types int:int:string #### A masked pattern was here #### name default.test_table2_n4 - numFiles 16 - numRows 500 - rawDataSize 7218 - serialization.ddl struct test_table2_n4 { i32 key, i32 key2, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7718 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","key2":"true","value":"true"}} SORTBUCKETCOLSPREFIX TRUE bucket_count 16 bucket_field_name key @@ -1134,14 +1020,8 @@ STAGE PLANS: columns.types int:int:string #### A masked pattern was here #### name default.test_table2_n4 - numFiles 16 - numRows 500 - rawDataSize 7218 - serialization.ddl struct test_table2_n4 { i32 key, i32 key2, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7718 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table2_n4 name: default.test_table2_n4 diff --git a/ql/src/test/results/clientpositive/llap/stats0.q.out b/ql/src/test/results/clientpositive/llap/stats0.q.out index 695ed643ab..2b08b61c17 100644 --- a/ql/src/test/results/clientpositive/llap/stats0.q.out +++ b/ql/src/test/results/clientpositive/llap/stats0.q.out @@ -56,8 +56,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -65,14 +63,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.stats_non_partitioned - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct stats_non_partitioned { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.stats_non_partitioned TotalFiles: 1 @@ -108,30 +100,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -139,14 +121,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -198,8 +174,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -207,14 +181,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.stats_non_partitioned - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct stats_non_partitioned { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.stats_non_partitioned @@ -1506,8 +1474,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1515,14 +1481,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.stats_non_partitioned - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct stats_non_partitioned { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.stats_non_partitioned TotalFiles: 1 @@ -1558,30 +1518,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1589,14 +1539,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -1648,8 +1592,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -1657,14 +1599,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.stats_non_partitioned - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct stats_non_partitioned { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.stats_non_partitioned diff --git a/ql/src/test/results/clientpositive/llap/stats11.q.out b/ql/src/test/results/clientpositive/llap/stats11.q.out index 71a1d9da15..b69b8e2c79 100644 --- a/ql/src/test/results/clientpositive/llap/stats11.q.out +++ b/ql/src/test/results/clientpositive/llap/stats11.q.out @@ -371,18 +371,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n15 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n15 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -397,14 +390,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n15 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n15 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_n15 name: default.srcbucket_mapjoin_n15 @@ -453,20 +440,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n16 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n16 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -483,10 +463,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n16 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n16 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n16 name: default.srcbucket_mapjoin_part_n16 @@ -521,8 +499,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -530,14 +506,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n7 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n7 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n7 TotalFiles: 1 @@ -608,8 +578,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -617,14 +585,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n7 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct bucketmapjoin_tmp_result_n7 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n7 @@ -823,18 +785,11 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n15 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n15 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -849,14 +804,8 @@ STAGE PLANS: columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_n15 - numFiles 2 - numRows 0 - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_n15 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2750 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_n15 name: default.srcbucket_mapjoin_n15 @@ -905,20 +854,13 @@ STAGE PLANS: bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_n16 - numFiles 4 - numRows 0 partition_columns ds partition_columns.types string - rawDataSize 0 - serialization.ddl struct srcbucket_mapjoin_part_n16 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -935,10 +877,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_n16 partition_columns ds partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_n16 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_n16 name: default.srcbucket_mapjoin_part_n16 @@ -973,8 +913,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -982,14 +920,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n7 - numFiles 1 - numRows 464 - rawDataSize 8519 - serialization.ddl struct bucketmapjoin_tmp_result_n7 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 8983 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n7 TotalFiles: 1 @@ -1060,8 +992,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value1":"true","value2":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value1,value2 @@ -1069,14 +999,8 @@ STAGE PLANS: columns.types string:string:string #### A masked pattern was here #### name default.bucketmapjoin_tmp_result_n7 - numFiles 1 - numRows 464 - rawDataSize 8519 - serialization.ddl struct bucketmapjoin_tmp_result_n7 { string key, string value1, string value2} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 8983 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.bucketmapjoin_tmp_result_n7 diff --git a/ql/src/test/results/clientpositive/llap/stats12.q.out b/ql/src/test/results/clientpositive/llap/stats12.q.out index b82bb0bfcd..e160da1d54 100644 --- a/ql/src/test/results/clientpositive/llap/stats12.q.out +++ b/ql/src/test/results/clientpositive/llap/stats12.q.out @@ -80,27 +80,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.analyze_srcpart_n3 - numFiles 1 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct analyze_srcpart_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -110,10 +103,8 @@ STAGE PLANS: name default.analyze_srcpart_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct analyze_srcpart_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.analyze_srcpart_n3 name: default.analyze_srcpart_n3 @@ -126,27 +117,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.analyze_srcpart_n3 - numFiles 1 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct analyze_srcpart_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -156,10 +140,8 @@ STAGE PLANS: name default.analyze_srcpart_n3 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct analyze_srcpart_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.analyze_srcpart_n3 name: default.analyze_srcpart_n3 diff --git a/ql/src/test/results/clientpositive/llap/stats13.q.out b/ql/src/test/results/clientpositive/llap/stats13.q.out index 6954cbd0b1..efd2b256d3 100644 --- a/ql/src/test/results/clientpositive/llap/stats13.q.out +++ b/ql/src/test/results/clientpositive/llap/stats13.q.out @@ -76,27 +76,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.analyze_srcpart - numFiles 1 partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct analyze_srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -106,10 +99,8 @@ STAGE PLANS: name default.analyze_srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct analyze_srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.analyze_srcpart name: default.analyze_srcpart diff --git a/ql/src/test/results/clientpositive/llap/stats3.q.out b/ql/src/test/results/clientpositive/llap/stats3.q.out index 8128d73def..3ef3c6e461 100644 --- a/ql/src/test/results/clientpositive/llap/stats3.q.out +++ b/ql/src/test/results/clientpositive/llap/stats3.q.out @@ -38,8 +38,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col1":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns col1 @@ -47,14 +45,8 @@ STAGE PLANS: columns.types string #### A masked pattern was here #### name default.hive_test_src_n3 - numFiles 0 - numRows 0 - rawDataSize 0 - serialization.ddl struct hive_test_src_n3 { string col1} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.hive_test_src_n3 diff --git a/ql/src/test/results/clientpositive/llap/temp_table_alter_partition_coltype.q.out b/ql/src/test/results/clientpositive/llap/temp_table_alter_partition_coltype.q.out index ead9709817..ea8d982154 100644 --- a/ql/src/test/results/clientpositive/llap/temp_table_alter_partition_coltype.q.out +++ b/ql/src/test/results/clientpositive/llap/temp_table_alter_partition_coltype.q.out @@ -193,30 +193,20 @@ STAGE PLANS: dt 100 ts 3.0 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.alter_coltype_temp - numFiles 1 - numRows 25 partition_columns dt/ts partition_columns.types string:double - rawDataSize 191 - serialization.ddl struct alter_coltype_temp { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -224,15 +214,10 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.alter_coltype_temp - numFiles 0 - numRows 0 partition_columns dt/ts partition_columns.types string:double - rawDataSize 0 - serialization.ddl struct alter_coltype_temp { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.alter_coltype_temp name: default.alter_coltype_temp @@ -245,30 +230,20 @@ STAGE PLANS: dt 100 ts 6.30 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.alter_coltype_temp - numFiles 1 - numRows 25 partition_columns dt/ts partition_columns.types string:double - rawDataSize 191 - serialization.ddl struct alter_coltype_temp { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -276,15 +251,10 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.alter_coltype_temp - numFiles 0 - numRows 0 partition_columns dt/ts partition_columns.types string:double - rawDataSize 0 - serialization.ddl struct alter_coltype_temp { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.alter_coltype_temp name: default.alter_coltype_temp @@ -412,30 +382,20 @@ STAGE PLANS: dt 100 ts 3.0 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.alter_coltype_temp - numFiles 1 - numRows 25 partition_columns dt/ts partition_columns.types string:double - rawDataSize 191 - serialization.ddl struct alter_coltype_temp { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -443,15 +403,10 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.alter_coltype_temp - numFiles 0 - numRows 0 partition_columns dt/ts partition_columns.types string:double - rawDataSize 0 - serialization.ddl struct alter_coltype_temp { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.alter_coltype_temp name: default.alter_coltype_temp @@ -464,30 +419,20 @@ STAGE PLANS: dt 100 ts 6.30 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.alter_coltype_temp - numFiles 1 - numRows 25 partition_columns dt/ts partition_columns.types string:double - rawDataSize 191 - serialization.ddl struct alter_coltype_temp { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -495,15 +440,10 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.alter_coltype_temp - numFiles 0 - numRows 0 partition_columns dt/ts partition_columns.types string:double - rawDataSize 0 - serialization.ddl struct alter_coltype_temp { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.alter_coltype_temp name: default.alter_coltype_temp @@ -641,30 +581,20 @@ STAGE PLANS: dt 100 ts 3.0 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.alter_coltype_temp - numFiles 1 - numRows 25 partition_columns dt/ts partition_columns.types string:double - rawDataSize 191 - serialization.ddl struct alter_coltype_temp { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -672,15 +602,10 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.alter_coltype_temp - numFiles 0 - numRows 0 partition_columns dt/ts partition_columns.types string:double - rawDataSize 0 - serialization.ddl struct alter_coltype_temp { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.alter_coltype_temp name: default.alter_coltype_temp @@ -691,30 +616,20 @@ STAGE PLANS: dt 100 ts 6.30 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.alter_coltype_temp - numFiles 1 - numRows 25 partition_columns dt/ts partition_columns.types string:double - rawDataSize 191 - serialization.ddl struct alter_coltype_temp { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 216 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -722,15 +637,10 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.alter_coltype_temp - numFiles 0 - numRows 0 partition_columns dt/ts partition_columns.types string:double - rawDataSize 0 - serialization.ddl struct alter_coltype_temp { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.alter_coltype_temp name: default.alter_coltype_temp @@ -895,30 +805,20 @@ STAGE PLANS: partcol1 1 partcol2 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 column.name.delimiter , columns intcol - columns.comments columns.types string #### A masked pattern was here #### name pt.alterdynamic_part_table_temp - numFiles 2 - numRows 2 partition_columns partcol1/partcol2 partition_columns.types int:string - rawDataSize 3 - serialization.ddl struct alterdynamic_part_table_temp { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"intcol":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns intcol @@ -926,15 +826,10 @@ STAGE PLANS: columns.types string #### A masked pattern was here #### name pt.alterdynamic_part_table_temp - numFiles 0 - numRows 0 partition_columns partcol1/partcol2 partition_columns.types int:string - rawDataSize 0 - serialization.ddl struct alterdynamic_part_table_temp { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: pt.alterdynamic_part_table_temp name: pt.alterdynamic_part_table_temp @@ -976,30 +871,20 @@ STAGE PLANS: partcol1 2 partcol2 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true"} - bucket_count -1 column.name.delimiter , columns intcol - columns.comments columns.types string #### A masked pattern was here #### name pt.alterdynamic_part_table_temp - numFiles 1 - numRows 1 partition_columns partcol1/partcol2 partition_columns.types int:string - rawDataSize 1 - serialization.ddl struct alterdynamic_part_table_temp { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"intcol":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns intcol @@ -1007,15 +892,10 @@ STAGE PLANS: columns.types string #### A masked pattern was here #### name pt.alterdynamic_part_table_temp - numFiles 0 - numRows 0 partition_columns partcol1/partcol2 partition_columns.types int:string - rawDataSize 0 - serialization.ddl struct alterdynamic_part_table_temp { string intcol} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: pt.alterdynamic_part_table_temp name: pt.alterdynamic_part_table_temp diff --git a/ql/src/test/results/clientpositive/llap/temp_table_display_colstats_tbllvl.q.out b/ql/src/test/results/clientpositive/llap/temp_table_display_colstats_tbllvl.q.out index 29fb49bdbd..e52adde9cd 100644 --- a/ql/src/test/results/clientpositive/llap/temp_table_display_colstats_tbllvl.q.out +++ b/ql/src/test/results/clientpositive/llap/temp_table_display_colstats_tbllvl.q.out @@ -288,25 +288,18 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite - columns.comments columns.types string:string:string:float:string:string:string:string:int field.delim | #### A masked pattern was here #### name default.uservisits_web_text_none - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7060 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: EXTERNAL TRUE - bucket_count -1 bucketing_version 2 column.name.delimiter , columns sourceip,desturl,visitdate,adrevenue,useragent,ccode,lcode,skeyword,avgtimeonsite @@ -315,13 +308,8 @@ STAGE PLANS: field.delim | #### A masked pattern was here #### name default.uservisits_web_text_none - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct uservisits_web_text_none { string sourceip, string desturl, string visitdate, float adrevenue, string useragent, string ccode, string lcode, string skeyword, i32 avgtimeonsite} serialization.format | serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7060 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.uservisits_web_text_none name: default.uservisits_web_text_none diff --git a/ql/src/test/results/clientpositive/llap/tez_fixed_bucket_pruning.q.out b/ql/src/test/results/clientpositive/llap/tez_fixed_bucket_pruning.q.out index bbb7d37fee..19c69b8c87 100644 --- a/ql/src/test/results/clientpositive/llap/tez_fixed_bucket_pruning.q.out +++ b/ql/src/test/results/clientpositive/llap/tez_fixed_bucket_pruning.q.out @@ -579,30 +579,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"business_dept_object_id":"true","business_partner_percentage":"true","charge_code_object_id":"true","date_key":"true","fy_month_key":"true","fy_quarter_key":"true","fy_year_key":"true","month_key":"true","plan_detail_object_id":"true","project_object_id":"true","quarter_key":"true","resource_object_id":"true","slice_date":"true","split_amount":"true","split_units":"true","supplier_object_id":"true","transclass_object_id":"true","week_key":"true","year_key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns plan_detail_object_id,project_object_id,charge_code_object_id,transclass_object_id,resource_object_id,slice_date,split_amount,split_units,year_key,quarter_key,month_key,week_key,date_key,fy_year_key,fy_quarter_key,fy_month_key,supplier_object_id,business_dept_object_id,business_partner_percentage - columns.comments columns.types bigint:bigint:bigint:bigint:bigint:varchar(50):varchar(50):varchar(50):varchar(20):varchar(20):varchar(50):varchar(50):varchar(50):varchar(50):string:string:bigint:bigint:decimal(38,8) #### A masked pattern was here #### name default.l3_clarity__l3_monthly_dw_factplan_dw_stg_2018022300104_1 - numFiles 1 - numRows 15 - rawDataSize 16430 - serialization.ddl struct l3_clarity__l3_monthly_dw_factplan_dw_stg_2018022300104_1 { i64 plan_detail_object_id, i64 project_object_id, i64 charge_code_object_id, i64 transclass_object_id, i64 resource_object_id, varchar(50) slice_date, varchar(50) split_amount, varchar(50) split_units, varchar(20) year_key, varchar(20) quarter_key, varchar(50) month_key, varchar(50) week_key, varchar(50) date_key, varchar(50) fy_year_key, string fy_quarter_key, string fy_month_key, i64 supplier_object_id, i64 business_dept_object_id, decimal(38,8) business_partner_percentage} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 3483 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"business_dept_object_id":"true","business_partner_percentage":"true","charge_code_object_id":"true","date_key":"true","fy_month_key":"true","fy_quarter_key":"true","fy_year_key":"true","month_key":"true","plan_detail_object_id":"true","project_object_id":"true","quarter_key":"true","resource_object_id":"true","slice_date":"true","split_amount":"true","split_units":"true","supplier_object_id":"true","transclass_object_id":"true","week_key":"true","year_key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns plan_detail_object_id,project_object_id,charge_code_object_id,transclass_object_id,resource_object_id,slice_date,split_amount,split_units,year_key,quarter_key,month_key,week_key,date_key,fy_year_key,fy_quarter_key,fy_month_key,supplier_object_id,business_dept_object_id,business_partner_percentage @@ -610,14 +600,8 @@ STAGE PLANS: columns.types bigint:bigint:bigint:bigint:bigint:varchar(50):varchar(50):varchar(50):varchar(20):varchar(20):varchar(50):varchar(50):varchar(50):varchar(50):string:string:bigint:bigint:decimal(38,8) #### A masked pattern was here #### name default.l3_clarity__l3_monthly_dw_factplan_dw_stg_2018022300104_1 - numFiles 1 - numRows 15 - rawDataSize 16430 - serialization.ddl struct l3_clarity__l3_monthly_dw_factplan_dw_stg_2018022300104_1 { i64 plan_detail_object_id, i64 project_object_id, i64 charge_code_object_id, i64 transclass_object_id, i64 resource_object_id, varchar(50) slice_date, varchar(50) split_amount, varchar(50) split_units, varchar(20) year_key, varchar(20) quarter_key, varchar(50) month_key, varchar(50) week_key, varchar(50) date_key, varchar(50) fy_year_key, string fy_quarter_key, string fy_month_key, i64 supplier_object_id, i64 business_dept_object_id, decimal(38,8) business_partner_percentage} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 3483 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.l3_clarity__l3_monthly_dw_factplan_dw_stg_2018022300104_1 name: default.l3_clarity__l3_monthly_dw_factplan_dw_stg_2018022300104_1 @@ -653,30 +637,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"l3_snapshot_number":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns l3_snapshot_number - columns.comments columns.types bigint #### A masked pattern was here #### name default.l3_clarity__l3_snap_number_2018022300104 - numFiles 1 - numRows 1 - rawDataSize 6 - serialization.ddl struct l3_clarity__l3_snap_number_2018022300104 { i64 l3_snapshot_number} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"l3_snapshot_number":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns l3_snapshot_number @@ -684,14 +658,8 @@ STAGE PLANS: columns.types bigint #### A masked pattern was here #### name default.l3_clarity__l3_snap_number_2018022300104 - numFiles 1 - numRows 1 - rawDataSize 6 - serialization.ddl struct l3_clarity__l3_snap_number_2018022300104 { i64 l3_snapshot_number} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.l3_clarity__l3_snap_number_2018022300104 name: default.l3_clarity__l3_snap_number_2018022300104 @@ -734,30 +702,21 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"bmo_cost_type":"true","bmo_fiscal_year":"true","charge_code_key":"true","charge_code_object_id":"true","clarity_updated_date":"true","finplan_detail_object_id":"true","idp_audit_id":"true","idp_data_date":"true","idp_warehouse_id":"true","is_latest_snapshot":"true","l3_created_date":"true","l3_snapshot_number":"true","last_updated_by":"true","latest_fiscal_budget_plan":"true","percentage":"true","period_end":"true","period_start":"true","period_type":"true","plan_category":"true","plan_code":"true","plan_description":"true","plan_key":"true","plan_name":"true","plan_of_record":"true","plan_status":"true","plan_type":"true","project_key":"true","project_object_id":"true","resoruce_object_id":"true","resource_key":"true","transclass_key":"true","txn_class_object_id":"true"}} bucket_count 64 bucket_field_name idp_data_date bucketing_version 2 column.name.delimiter , columns idp_warehouse_id,idp_audit_id,idp_data_date,l3_snapshot_number,plan_key,project_key,charge_code_key,transclass_key,resource_key,finplan_detail_object_id,project_object_id,txn_class_object_id,charge_code_object_id,resoruce_object_id,plan_name,plan_code,plan_type,period_type,plan_description,plan_status,period_start,period_end,plan_of_record,percentage,l3_created_date,bmo_cost_type,bmo_fiscal_year,clarity_updated_date,is_latest_snapshot,latest_fiscal_budget_plan,plan_category,last_updated_by - columns.comments columns.types bigint:bigint:date:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:varchar(1500):varchar(500):varchar(50):varchar(50):varchar(3000):varchar(50):varchar(50):varchar(50):varchar(1):decimal(32,6):timestamp:varchar(30):varchar(50):timestamp:bigint:bigint:varchar(70):varchar(250) #### A masked pattern was here #### name default.l3_monthly_dw_dimplan - numFiles 1 - numRows 180340 - rawDataSize 269826156 - serialization.ddl struct l3_monthly_dw_dimplan { i64 idp_warehouse_id, i64 idp_audit_id, date idp_data_date, i64 l3_snapshot_number, i64 plan_key, i64 project_key, i64 charge_code_key, i64 transclass_key, i64 resource_key, i64 finplan_detail_object_id, i64 project_object_id, i64 txn_class_object_id, i64 charge_code_object_id, i64 resoruce_object_id, varchar(1500) plan_name, varchar(500) plan_code, varchar(50) plan_type, varchar(50) period_type, varchar(3000) plan_description, varchar(50) plan_status, varchar(50) period_start, varchar(50) period_end, varchar(1) plan_of_record, decimal(32,6) percentage, timestamp l3_created_date, varchar(30) bmo_cost_type, varchar(50) bmo_fiscal_year, timestamp clarity_updated_date, i64 is_latest_snapshot, i64 latest_fiscal_budget_plan, varchar(70) plan_category, varchar(250) last_updated_by} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 5242699 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"bmo_cost_type":"true","bmo_fiscal_year":"true","charge_code_key":"true","charge_code_object_id":"true","clarity_updated_date":"true","finplan_detail_object_id":"true","idp_audit_id":"true","idp_data_date":"true","idp_warehouse_id":"true","is_latest_snapshot":"true","l3_created_date":"true","l3_snapshot_number":"true","last_updated_by":"true","latest_fiscal_budget_plan":"true","percentage":"true","period_end":"true","period_start":"true","period_type":"true","plan_category":"true","plan_code":"true","plan_description":"true","plan_key":"true","plan_name":"true","plan_of_record":"true","plan_status":"true","plan_type":"true","project_key":"true","project_object_id":"true","resoruce_object_id":"true","resource_key":"true","transclass_key":"true","txn_class_object_id":"true"}} bucket_count 64 bucket_field_name idp_data_date bucketing_version 2 @@ -767,14 +726,8 @@ STAGE PLANS: columns.types bigint:bigint:date:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:varchar(1500):varchar(500):varchar(50):varchar(50):varchar(3000):varchar(50):varchar(50):varchar(50):varchar(1):decimal(32,6):timestamp:varchar(30):varchar(50):timestamp:bigint:bigint:varchar(70):varchar(250) #### A masked pattern was here #### name default.l3_monthly_dw_dimplan - numFiles 1 - numRows 180340 - rawDataSize 269826156 - serialization.ddl struct l3_monthly_dw_dimplan { i64 idp_warehouse_id, i64 idp_audit_id, date idp_data_date, i64 l3_snapshot_number, i64 plan_key, i64 project_key, i64 charge_code_key, i64 transclass_key, i64 resource_key, i64 finplan_detail_object_id, i64 project_object_id, i64 txn_class_object_id, i64 charge_code_object_id, i64 resoruce_object_id, varchar(1500) plan_name, varchar(500) plan_code, varchar(50) plan_type, varchar(50) period_type, varchar(3000) plan_description, varchar(50) plan_status, varchar(50) period_start, varchar(50) period_end, varchar(1) plan_of_record, decimal(32,6) percentage, timestamp l3_created_date, varchar(30) bmo_cost_type, varchar(50) bmo_fiscal_year, timestamp clarity_updated_date, i64 is_latest_snapshot, i64 latest_fiscal_budget_plan, varchar(70) plan_category, varchar(250) last_updated_by} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 5242699 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.l3_monthly_dw_dimplan name: default.l3_monthly_dw_dimplan @@ -817,30 +770,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"idp_data_date":"true","l3_created_date":"true","l3_snapshot_number":"true","project_key":"true","project_object_id":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns project_key,l3_snapshot_number,l3_created_date,project_object_id,idp_data_date - columns.comments columns.types bigint:bigint:timestamp:bigint:date #### A masked pattern was here #### name default.l3_clarity__l3_monthly_dw_factplan_datajoin_1_s2_2018022300104_1 - numFiles 1 - numRows 1 - rawDataSize 120 - serialization.ddl struct l3_clarity__l3_monthly_dw_factplan_datajoin_1_s2_2018022300104_1 { i64 project_key, i64 l3_snapshot_number, timestamp l3_created_date, i64 project_object_id, date idp_data_date} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 677 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"idp_data_date":"true","l3_created_date":"true","l3_snapshot_number":"true","project_key":"true","project_object_id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns project_key,l3_snapshot_number,l3_created_date,project_object_id,idp_data_date @@ -848,14 +791,8 @@ STAGE PLANS: columns.types bigint:bigint:timestamp:bigint:date #### A masked pattern was here #### name default.l3_clarity__l3_monthly_dw_factplan_datajoin_1_s2_2018022300104_1 - numFiles 1 - numRows 1 - rawDataSize 120 - serialization.ddl struct l3_clarity__l3_monthly_dw_factplan_datajoin_1_s2_2018022300104_1 { i64 project_key, i64 l3_snapshot_number, timestamp l3_created_date, i64 project_object_id, date idp_data_date} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 677 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.l3_clarity__l3_monthly_dw_factplan_datajoin_1_s2_2018022300104_1 name: default.l3_clarity__l3_monthly_dw_factplan_datajoin_1_s2_2018022300104_1 @@ -1104,30 +1041,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"business_dept_object_id":"true","business_partner_percentage":"true","charge_code_object_id":"true","date_key":"true","fy_month_key":"true","fy_quarter_key":"true","fy_year_key":"true","month_key":"true","plan_detail_object_id":"true","project_object_id":"true","quarter_key":"true","resource_object_id":"true","slice_date":"true","split_amount":"true","split_units":"true","supplier_object_id":"true","transclass_object_id":"true","week_key":"true","year_key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns plan_detail_object_id,project_object_id,charge_code_object_id,transclass_object_id,resource_object_id,slice_date,split_amount,split_units,year_key,quarter_key,month_key,week_key,date_key,fy_year_key,fy_quarter_key,fy_month_key,supplier_object_id,business_dept_object_id,business_partner_percentage - columns.comments columns.types bigint:bigint:bigint:bigint:bigint:varchar(50):varchar(50):varchar(50):varchar(20):varchar(20):varchar(50):varchar(50):varchar(50):varchar(50):string:string:bigint:bigint:decimal(38,8) #### A masked pattern was here #### name default.l3_clarity__l3_monthly_dw_factplan_dw_stg_2018022300104_1 - numFiles 1 - numRows 15 - rawDataSize 16430 - serialization.ddl struct l3_clarity__l3_monthly_dw_factplan_dw_stg_2018022300104_1 { i64 plan_detail_object_id, i64 project_object_id, i64 charge_code_object_id, i64 transclass_object_id, i64 resource_object_id, varchar(50) slice_date, varchar(50) split_amount, varchar(50) split_units, varchar(20) year_key, varchar(20) quarter_key, varchar(50) month_key, varchar(50) week_key, varchar(50) date_key, varchar(50) fy_year_key, string fy_quarter_key, string fy_month_key, i64 supplier_object_id, i64 business_dept_object_id, decimal(38,8) business_partner_percentage} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 3483 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"business_dept_object_id":"true","business_partner_percentage":"true","charge_code_object_id":"true","date_key":"true","fy_month_key":"true","fy_quarter_key":"true","fy_year_key":"true","month_key":"true","plan_detail_object_id":"true","project_object_id":"true","quarter_key":"true","resource_object_id":"true","slice_date":"true","split_amount":"true","split_units":"true","supplier_object_id":"true","transclass_object_id":"true","week_key":"true","year_key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns plan_detail_object_id,project_object_id,charge_code_object_id,transclass_object_id,resource_object_id,slice_date,split_amount,split_units,year_key,quarter_key,month_key,week_key,date_key,fy_year_key,fy_quarter_key,fy_month_key,supplier_object_id,business_dept_object_id,business_partner_percentage @@ -1135,14 +1062,8 @@ STAGE PLANS: columns.types bigint:bigint:bigint:bigint:bigint:varchar(50):varchar(50):varchar(50):varchar(20):varchar(20):varchar(50):varchar(50):varchar(50):varchar(50):string:string:bigint:bigint:decimal(38,8) #### A masked pattern was here #### name default.l3_clarity__l3_monthly_dw_factplan_dw_stg_2018022300104_1 - numFiles 1 - numRows 15 - rawDataSize 16430 - serialization.ddl struct l3_clarity__l3_monthly_dw_factplan_dw_stg_2018022300104_1 { i64 plan_detail_object_id, i64 project_object_id, i64 charge_code_object_id, i64 transclass_object_id, i64 resource_object_id, varchar(50) slice_date, varchar(50) split_amount, varchar(50) split_units, varchar(20) year_key, varchar(20) quarter_key, varchar(50) month_key, varchar(50) week_key, varchar(50) date_key, varchar(50) fy_year_key, string fy_quarter_key, string fy_month_key, i64 supplier_object_id, i64 business_dept_object_id, decimal(38,8) business_partner_percentage} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 3483 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.l3_clarity__l3_monthly_dw_factplan_dw_stg_2018022300104_1 name: default.l3_clarity__l3_monthly_dw_factplan_dw_stg_2018022300104_1 @@ -1178,30 +1099,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"l3_snapshot_number":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns l3_snapshot_number - columns.comments columns.types bigint #### A masked pattern was here #### name default.l3_clarity__l3_snap_number_2018022300104 - numFiles 1 - numRows 1 - rawDataSize 6 - serialization.ddl struct l3_clarity__l3_snap_number_2018022300104 { i64 l3_snapshot_number} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"l3_snapshot_number":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns l3_snapshot_number @@ -1209,14 +1120,8 @@ STAGE PLANS: columns.types bigint #### A masked pattern was here #### name default.l3_clarity__l3_snap_number_2018022300104 - numFiles 1 - numRows 1 - rawDataSize 6 - serialization.ddl struct l3_clarity__l3_snap_number_2018022300104 { i64 l3_snapshot_number} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.l3_clarity__l3_snap_number_2018022300104 name: default.l3_clarity__l3_snap_number_2018022300104 @@ -1259,30 +1164,21 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"bmo_cost_type":"true","bmo_fiscal_year":"true","charge_code_key":"true","charge_code_object_id":"true","clarity_updated_date":"true","finplan_detail_object_id":"true","idp_audit_id":"true","idp_data_date":"true","idp_warehouse_id":"true","is_latest_snapshot":"true","l3_created_date":"true","l3_snapshot_number":"true","last_updated_by":"true","latest_fiscal_budget_plan":"true","percentage":"true","period_end":"true","period_start":"true","period_type":"true","plan_category":"true","plan_code":"true","plan_description":"true","plan_key":"true","plan_name":"true","plan_of_record":"true","plan_status":"true","plan_type":"true","project_key":"true","project_object_id":"true","resoruce_object_id":"true","resource_key":"true","transclass_key":"true","txn_class_object_id":"true"}} bucket_count 64 bucket_field_name idp_data_date bucketing_version 2 column.name.delimiter , columns idp_warehouse_id,idp_audit_id,idp_data_date,l3_snapshot_number,plan_key,project_key,charge_code_key,transclass_key,resource_key,finplan_detail_object_id,project_object_id,txn_class_object_id,charge_code_object_id,resoruce_object_id,plan_name,plan_code,plan_type,period_type,plan_description,plan_status,period_start,period_end,plan_of_record,percentage,l3_created_date,bmo_cost_type,bmo_fiscal_year,clarity_updated_date,is_latest_snapshot,latest_fiscal_budget_plan,plan_category,last_updated_by - columns.comments columns.types bigint:bigint:date:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:varchar(1500):varchar(500):varchar(50):varchar(50):varchar(3000):varchar(50):varchar(50):varchar(50):varchar(1):decimal(32,6):timestamp:varchar(30):varchar(50):timestamp:bigint:bigint:varchar(70):varchar(250) #### A masked pattern was here #### name default.l3_monthly_dw_dimplan - numFiles 1 - numRows 180340 - rawDataSize 269826156 - serialization.ddl struct l3_monthly_dw_dimplan { i64 idp_warehouse_id, i64 idp_audit_id, date idp_data_date, i64 l3_snapshot_number, i64 plan_key, i64 project_key, i64 charge_code_key, i64 transclass_key, i64 resource_key, i64 finplan_detail_object_id, i64 project_object_id, i64 txn_class_object_id, i64 charge_code_object_id, i64 resoruce_object_id, varchar(1500) plan_name, varchar(500) plan_code, varchar(50) plan_type, varchar(50) period_type, varchar(3000) plan_description, varchar(50) plan_status, varchar(50) period_start, varchar(50) period_end, varchar(1) plan_of_record, decimal(32,6) percentage, timestamp l3_created_date, varchar(30) bmo_cost_type, varchar(50) bmo_fiscal_year, timestamp clarity_updated_date, i64 is_latest_snapshot, i64 latest_fiscal_budget_plan, varchar(70) plan_category, varchar(250) last_updated_by} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 5242699 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"bmo_cost_type":"true","bmo_fiscal_year":"true","charge_code_key":"true","charge_code_object_id":"true","clarity_updated_date":"true","finplan_detail_object_id":"true","idp_audit_id":"true","idp_data_date":"true","idp_warehouse_id":"true","is_latest_snapshot":"true","l3_created_date":"true","l3_snapshot_number":"true","last_updated_by":"true","latest_fiscal_budget_plan":"true","percentage":"true","period_end":"true","period_start":"true","period_type":"true","plan_category":"true","plan_code":"true","plan_description":"true","plan_key":"true","plan_name":"true","plan_of_record":"true","plan_status":"true","plan_type":"true","project_key":"true","project_object_id":"true","resoruce_object_id":"true","resource_key":"true","transclass_key":"true","txn_class_object_id":"true"}} bucket_count 64 bucket_field_name idp_data_date bucketing_version 2 @@ -1292,14 +1188,8 @@ STAGE PLANS: columns.types bigint:bigint:date:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:bigint:varchar(1500):varchar(500):varchar(50):varchar(50):varchar(3000):varchar(50):varchar(50):varchar(50):varchar(1):decimal(32,6):timestamp:varchar(30):varchar(50):timestamp:bigint:bigint:varchar(70):varchar(250) #### A masked pattern was here #### name default.l3_monthly_dw_dimplan - numFiles 1 - numRows 180340 - rawDataSize 269826156 - serialization.ddl struct l3_monthly_dw_dimplan { i64 idp_warehouse_id, i64 idp_audit_id, date idp_data_date, i64 l3_snapshot_number, i64 plan_key, i64 project_key, i64 charge_code_key, i64 transclass_key, i64 resource_key, i64 finplan_detail_object_id, i64 project_object_id, i64 txn_class_object_id, i64 charge_code_object_id, i64 resoruce_object_id, varchar(1500) plan_name, varchar(500) plan_code, varchar(50) plan_type, varchar(50) period_type, varchar(3000) plan_description, varchar(50) plan_status, varchar(50) period_start, varchar(50) period_end, varchar(1) plan_of_record, decimal(32,6) percentage, timestamp l3_created_date, varchar(30) bmo_cost_type, varchar(50) bmo_fiscal_year, timestamp clarity_updated_date, i64 is_latest_snapshot, i64 latest_fiscal_budget_plan, varchar(70) plan_category, varchar(250) last_updated_by} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 5242699 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.l3_monthly_dw_dimplan name: default.l3_monthly_dw_dimplan @@ -1342,30 +1232,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"idp_data_date":"true","l3_created_date":"true","l3_snapshot_number":"true","project_key":"true","project_object_id":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns project_key,l3_snapshot_number,l3_created_date,project_object_id,idp_data_date - columns.comments columns.types bigint:bigint:timestamp:bigint:date #### A masked pattern was here #### name default.l3_clarity__l3_monthly_dw_factplan_datajoin_1_s2_2018022300104_1 - numFiles 1 - numRows 1 - rawDataSize 120 - serialization.ddl struct l3_clarity__l3_monthly_dw_factplan_datajoin_1_s2_2018022300104_1 { i64 project_key, i64 l3_snapshot_number, timestamp l3_created_date, i64 project_object_id, date idp_data_date} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 677 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"idp_data_date":"true","l3_created_date":"true","l3_snapshot_number":"true","project_key":"true","project_object_id":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns project_key,l3_snapshot_number,l3_created_date,project_object_id,idp_data_date @@ -1373,14 +1253,8 @@ STAGE PLANS: columns.types bigint:bigint:timestamp:bigint:date #### A masked pattern was here #### name default.l3_clarity__l3_monthly_dw_factplan_datajoin_1_s2_2018022300104_1 - numFiles 1 - numRows 1 - rawDataSize 120 - serialization.ddl struct l3_clarity__l3_monthly_dw_factplan_datajoin_1_s2_2018022300104_1 { i64 project_key, i64 l3_snapshot_number, timestamp l3_created_date, i64 project_object_id, date idp_data_date} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 677 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.l3_clarity__l3_monthly_dw_factplan_datajoin_1_s2_2018022300104_1 name: default.l3_clarity__l3_monthly_dw_factplan_datajoin_1_s2_2018022300104_1 @@ -1567,30 +1441,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col_1":"true","col_2":"true","col_3":"true"}} bucket_count 4 bucket_field_name col_1 bucketing_version 2 column.name.delimiter , columns col_1,col_2,col_3 - columns.comments columns.types int:string:string #### A masked pattern was here #### name default.test_table - numFiles 3 - numRows 4 - rawDataSize 42 - serialization.ddl struct test_table { i32 col_1, string col_2, string col_3} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 46 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col_1":"true","col_2":"true","col_3":"true"}} bucket_count 4 bucket_field_name col_1 bucketing_version 2 @@ -1600,14 +1465,8 @@ STAGE PLANS: columns.types int:string:string #### A masked pattern was here #### name default.test_table - numFiles 3 - numRows 4 - rawDataSize 42 - serialization.ddl struct test_table { i32 col_1, string col_2, string col_3} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 46 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table name: default.test_table @@ -1725,30 +1584,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col_1":"true","col_2":"true","col_3":"true"}} bucket_count 4 bucket_field_name col_1 bucketing_version 2 column.name.delimiter , columns col_1,col_2,col_3 - columns.comments columns.types int:string:string #### A masked pattern was here #### name default.test_table - numFiles 3 - numRows 4 - rawDataSize 42 - serialization.ddl struct test_table { i32 col_1, string col_2, string col_3} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 46 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"col_1":"true","col_2":"true","col_3":"true"}} bucket_count 4 bucket_field_name col_1 bucketing_version 2 @@ -1758,14 +1608,8 @@ STAGE PLANS: columns.types int:string:string #### A masked pattern was here #### name default.test_table - numFiles 3 - numRows 4 - rawDataSize 42 - serialization.ddl struct test_table { i32 col_1, string col_2, string col_3} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 46 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.test_table name: default.test_table diff --git a/ql/src/test/results/clientpositive/llap/tez_join_result_complex.q.out b/ql/src/test/results/clientpositive/llap/tez_join_result_complex.q.out index 0b042b92d0..8acf2a02b4 100644 --- a/ql/src/test/results/clientpositive/llap/tez_join_result_complex.q.out +++ b/ql/src/test/results/clientpositive/llap/tez_join_result_complex.q.out @@ -219,7 +219,6 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns contact_event_id,ce_create_dt,ce_end_dt,contact_type,cnctevs_cd,contact_mode,cntvnst_stts_cd,total_transfers,ce_notes - columns.comments columns.types string:string:string:string:string:string:string:int:array field.delim #### A masked pattern was here #### @@ -227,20 +226,13 @@ STAGE PLANS: #### A masked pattern was here #### name default.ct_events_clean - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct ct_events_clean { string contact_event_id, string ce_create_dt, string ce_end_dt, string contact_type, string cnctevs_cd, string contact_mode, string cntvnst_stts_cd, i32 total_transfers, list ce_notes} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7501 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns contact_event_id,ce_create_dt,ce_end_dt,contact_type,cnctevs_cd,contact_mode,cntvnst_stts_cd,total_transfers,ce_notes @@ -252,14 +244,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.ct_events_clean - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct ct_events_clean { string contact_event_id, string ce_create_dt, string ce_end_dt, string contact_type, string cnctevs_cd, string contact_mode, string cntvnst_stts_cd, i32 total_transfers, list ce_notes} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7501 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.ct_events_clean name: default.ct_events_clean @@ -334,7 +320,6 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns cnctevn_id,svcrqst_id,svcrqst_crt_dts,subject_seq_no,plan_component,cust_segment,cnctyp_cd,cnctmd_cd,cnctevs_cd,svcrtyp_cd,svrstyp_cd,cmpltyp_cd,catsrsn_cd,apealvl_cd,cnstnty_cd,svcrqst_asrqst_ind,svcrqst_rtnorig_in,svcrqst_vwasof_dt,sum_reason_cd,sum_reason,crsr_master_claim_index,svcrqct_cds,svcrqst_lupdt,crsr_lupdt,cntevsds_lupdt,ignore_me,notes - columns.comments columns.types string:string:string:int:string:string:string:string:string:string:string:string:string:string:string:string:string:string:string:string:string:array:string:timestamp:string:int:array field.delim #### A masked pattern was here #### @@ -342,20 +327,13 @@ STAGE PLANS: #### A masked pattern was here #### name default.service_request_clean - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct service_request_clean { string cnctevn_id, string svcrqst_id, string svcrqst_crt_dts, i32 subject_seq_no, string plan_component, string cust_segment, string cnctyp_cd, string cnctmd_cd, string cnctevs_cd, string svcrtyp_cd, string svrstyp_cd, string cmpltyp_cd, string catsrsn_cd, string apealvl_cd, string cnstnty_cd, string svcrqst_asrqst_ind, string svcrqst_rtnorig_in, string svcrqst_vwasof_dt, string sum_reason_cd, string sum_reason, string crsr_master_claim_index, list svcrqct_cds, string svcrqst_lupdt, timestamp crsr_lupdt, string cntevsds_lupdt, i32 ignore_me, list notes} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 18162 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns cnctevn_id,svcrqst_id,svcrqst_crt_dts,subject_seq_no,plan_component,cust_segment,cnctyp_cd,cnctmd_cd,cnctevs_cd,svcrtyp_cd,svrstyp_cd,cmpltyp_cd,catsrsn_cd,apealvl_cd,cnstnty_cd,svcrqst_asrqst_ind,svcrqst_rtnorig_in,svcrqst_vwasof_dt,sum_reason_cd,sum_reason,crsr_master_claim_index,svcrqct_cds,svcrqst_lupdt,crsr_lupdt,cntevsds_lupdt,ignore_me,notes @@ -367,14 +345,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.service_request_clean - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct service_request_clean { string cnctevn_id, string svcrqst_id, string svcrqst_crt_dts, i32 subject_seq_no, string plan_component, string cust_segment, string cnctyp_cd, string cnctmd_cd, string cnctevs_cd, string svcrtyp_cd, string svrstyp_cd, string cmpltyp_cd, string catsrsn_cd, string apealvl_cd, string cnstnty_cd, string svcrqst_asrqst_ind, string svcrqst_rtnorig_in, string svcrqst_vwasof_dt, string sum_reason_cd, string sum_reason, string crsr_master_claim_index, list svcrqct_cds, string svcrqst_lupdt, timestamp crsr_lupdt, string cntevsds_lupdt, i32 ignore_me, list notes} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 18162 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.service_request_clean name: default.service_request_clean @@ -1228,7 +1200,6 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns contact_event_id,ce_create_dt,ce_end_dt,contact_type,cnctevs_cd,contact_mode,cntvnst_stts_cd,total_transfers,ce_notes - columns.comments columns.types string:string:string:string:string:string:string:int:array field.delim #### A masked pattern was here #### @@ -1236,20 +1207,13 @@ STAGE PLANS: #### A masked pattern was here #### name default.ct_events_clean - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct ct_events_clean { string contact_event_id, string ce_create_dt, string ce_end_dt, string contact_type, string cnctevs_cd, string contact_mode, string cntvnst_stts_cd, i32 total_transfers, list ce_notes} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7501 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns contact_event_id,ce_create_dt,ce_end_dt,contact_type,cnctevs_cd,contact_mode,cntvnst_stts_cd,total_transfers,ce_notes @@ -1261,14 +1225,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.ct_events_clean - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct ct_events_clean { string contact_event_id, string ce_create_dt, string ce_end_dt, string contact_type, string cnctevs_cd, string contact_mode, string cntvnst_stts_cd, i32 total_transfers, list ce_notes} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 7501 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.ct_events_clean name: default.ct_events_clean @@ -1343,7 +1301,6 @@ STAGE PLANS: bucketing_version 2 column.name.delimiter , columns cnctevn_id,svcrqst_id,svcrqst_crt_dts,subject_seq_no,plan_component,cust_segment,cnctyp_cd,cnctmd_cd,cnctevs_cd,svcrtyp_cd,svrstyp_cd,cmpltyp_cd,catsrsn_cd,apealvl_cd,cnstnty_cd,svcrqst_asrqst_ind,svcrqst_rtnorig_in,svcrqst_vwasof_dt,sum_reason_cd,sum_reason,crsr_master_claim_index,svcrqct_cds,svcrqst_lupdt,crsr_lupdt,cntevsds_lupdt,ignore_me,notes - columns.comments columns.types string:string:string:int:string:string:string:string:string:string:string:string:string:string:string:string:string:string:string:string:string:array:string:timestamp:string:int:array field.delim #### A masked pattern was here #### @@ -1351,20 +1308,13 @@ STAGE PLANS: #### A masked pattern was here #### name default.service_request_clean - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct service_request_clean { string cnctevn_id, string svcrqst_id, string svcrqst_crt_dts, i32 subject_seq_no, string plan_component, string cust_segment, string cnctyp_cd, string cnctmd_cd, string cnctevs_cd, string svcrtyp_cd, string svrstyp_cd, string cmpltyp_cd, string catsrsn_cd, string apealvl_cd, string cnstnty_cd, string svcrqst_asrqst_ind, string svcrqst_rtnorig_in, string svcrqst_vwasof_dt, string sum_reason_cd, string sum_reason, string crsr_master_claim_index, list svcrqct_cds, string svcrqst_lupdt, timestamp crsr_lupdt, string cntevsds_lupdt, i32 ignore_me, list notes} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 18162 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns cnctevn_id,svcrqst_id,svcrqst_crt_dts,subject_seq_no,plan_component,cust_segment,cnctyp_cd,cnctmd_cd,cnctevs_cd,svcrtyp_cd,svrstyp_cd,cmpltyp_cd,catsrsn_cd,apealvl_cd,cnstnty_cd,svcrqst_asrqst_ind,svcrqst_rtnorig_in,svcrqst_vwasof_dt,sum_reason_cd,sum_reason,crsr_master_claim_index,svcrqct_cds,svcrqst_lupdt,crsr_lupdt,cntevsds_lupdt,ignore_me,notes @@ -1376,14 +1326,8 @@ STAGE PLANS: #### A masked pattern was here #### name default.service_request_clean - numFiles 1 - numRows 0 - rawDataSize 0 - serialization.ddl struct service_request_clean { string cnctevn_id, string svcrqst_id, string svcrqst_crt_dts, i32 subject_seq_no, string plan_component, string cust_segment, string cnctyp_cd, string cnctmd_cd, string cnctevs_cd, string svcrtyp_cd, string svrstyp_cd, string cmpltyp_cd, string catsrsn_cd, string apealvl_cd, string cnstnty_cd, string svcrqst_asrqst_ind, string svcrqst_rtnorig_in, string svcrqst_vwasof_dt, string sum_reason_cd, string sum_reason, string crsr_master_claim_index, list svcrqct_cds, string svcrqst_lupdt, timestamp crsr_lupdt, string cntevsds_lupdt, i32 ignore_me, list notes} serialization.format serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 18162 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.service_request_clean name: default.service_request_clean diff --git a/ql/src/test/results/clientpositive/llap/topnkey_windowing.q.out b/ql/src/test/results/clientpositive/llap/topnkey_windowing.q.out index 6bf0dd418e..6f01a086be 100644 --- a/ql/src/test/results/clientpositive/llap/topnkey_windowing.q.out +++ b/ql/src/test/results/clientpositive/llap/topnkey_windowing.q.out @@ -428,30 +428,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"tw_code":"true","tw_value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns tw_code,tw_value - columns.comments columns.types string:double #### A masked pattern was here #### name default.topnkey_windowing - numFiles 1 - numRows 26 - rawDataSize 176 - serialization.ddl struct topnkey_windowing { string tw_code, double tw_value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 202 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"tw_code":"true","tw_value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns tw_code,tw_value @@ -459,14 +449,8 @@ STAGE PLANS: columns.types string:double #### A masked pattern was here #### name default.topnkey_windowing - numFiles 1 - numRows 26 - rawDataSize 176 - serialization.ddl struct topnkey_windowing { string tw_code, double tw_value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 202 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.topnkey_windowing name: default.topnkey_windowing @@ -633,30 +617,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"tw_code":"true","tw_value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns tw_code,tw_value - columns.comments columns.types string:double #### A masked pattern was here #### name default.topnkey_windowing - numFiles 1 - numRows 26 - rawDataSize 176 - serialization.ddl struct topnkey_windowing { string tw_code, double tw_value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 202 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"tw_code":"true","tw_value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns tw_code,tw_value @@ -664,14 +638,8 @@ STAGE PLANS: columns.types string:double #### A masked pattern was here #### name default.topnkey_windowing - numFiles 1 - numRows 26 - rawDataSize 176 - serialization.ddl struct topnkey_windowing { string tw_code, double tw_value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 202 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.topnkey_windowing name: default.topnkey_windowing diff --git a/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out b/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out index 1b5e505490..5a5396e10d 100644 --- a/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out +++ b/ql/src/test/results/clientpositive/llap/unionDistinct_1.q.out @@ -3939,31 +3939,21 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"k0":"true","k1":"true","k2":"true","k3":"true","k4":"true","k5":"true"}} - bucket_count -1 column.name.delimiter , columns k0,k1,k2,k3,k4,k5 - columns.comments columns.types string:string:string:string:string:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.dst_union22_delta_n0 - numFiles 1 - numRows 500 partition_columns ds partition_columns.types string - rawDataSize 16936 - serialization.ddl struct dst_union22_delta_n0 { string k0, string k1, string k2, string k3, string k4, string k5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17436 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k0,k1,k2,k3,k4,k5 @@ -3974,10 +3964,8 @@ STAGE PLANS: name default.dst_union22_delta_n0 partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22_delta_n0 { string k0, string k1, string k2, string k3, string k4, string k5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22_delta_n0 name: default.dst_union22_delta_n0 @@ -4047,31 +4035,21 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"k1":"true","k2":"true","k3":"true","k4":"true"}} - bucket_count -1 column.name.delimiter , columns k1,k2,k3,k4 - columns.comments columns.types string:string:string:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.dst_union22_n0 - numFiles 1 - numRows 500 partition_columns ds partition_columns.types string - rawDataSize 11124 - serialization.ddl struct dst_union22_n0 { string k1, string k2, string k3, string k4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k1,k2,k3,k4 @@ -4082,10 +4060,8 @@ STAGE PLANS: name default.dst_union22_n0 partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22_n0 { string k1, string k2, string k3, string k4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22_n0 name: default.dst_union22_n0 @@ -4129,31 +4105,21 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"k0":"true","k1":"true","k2":"true","k3":"true","k4":"true","k5":"true"}} - bucket_count -1 column.name.delimiter , columns k0,k1,k2,k3,k4,k5 - columns.comments columns.types string:string:string:string:string:string #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.dst_union22_delta_n0 - numFiles 1 - numRows 500 partition_columns ds partition_columns.types string - rawDataSize 16936 - serialization.ddl struct dst_union22_delta_n0 { string k0, string k1, string k2, string k3, string k4, string k5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17436 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k0,k1,k2,k3,k4,k5 @@ -4164,10 +4130,8 @@ STAGE PLANS: name default.dst_union22_delta_n0 partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22_delta_n0 { string k0, string k1, string k2, string k3, string k4, string k5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22_delta_n0 name: default.dst_union22_delta_n0 @@ -4195,7 +4159,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k1,k2,k3,k4 @@ -4206,10 +4169,8 @@ STAGE PLANS: name default.dst_union22_n0 partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22_n0 { string k1, string k2, string k3, string k4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22_n0 TotalFiles: 1 @@ -4292,7 +4253,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k1,k2,k3,k4 @@ -4303,10 +4263,8 @@ STAGE PLANS: name default.dst_union22_n0 partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22_n0 { string k1, string k2, string k3, string k4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22_n0 @@ -5287,31 +5245,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src2_n2 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src2_n2 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -5320,14 +5268,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src2_n2 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src2_n2 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src2_n2 name: default.src2_n2 @@ -5372,31 +5314,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src5_n1 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src5_n1 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -5405,14 +5337,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src5_n1 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src5_n1 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src5_n1 name: default.src5_n1 @@ -5459,31 +5385,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src3 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src3 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -5492,14 +5408,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src3 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src3 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src3 name: default.src3 @@ -5546,31 +5456,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src4 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src4 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -5579,14 +5479,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src4 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src4 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src4 name: default.src4 @@ -5856,31 +5750,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src2_n2 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src2_n2 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -5889,14 +5773,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src2_n2 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src2_n2 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src2_n2 name: default.src2_n2 @@ -5943,31 +5821,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src3 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src3 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -5976,14 +5844,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src3 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src3 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src3 name: default.src3 @@ -6024,31 +5886,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src4 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src4 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -6057,14 +5909,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src4 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src4 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src4 name: default.src4 @@ -6127,31 +5973,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src5_n1 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src5_n1 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -6160,14 +5996,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src5_n1 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src5_n1 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src5_n1 name: default.src5_n1 @@ -6380,31 +6210,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src2_n2 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src2_n2 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -6413,14 +6233,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src2_n2 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src2_n2 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src2_n2 name: default.src2_n2 @@ -6467,31 +6281,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src3 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src3 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -6500,14 +6304,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src3 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src3 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src3 name: default.src3 @@ -6568,31 +6366,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src4 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src4 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -6601,14 +6389,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src4 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src4 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src4 name: default.src4 @@ -6649,31 +6431,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src5_n1 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src5_n1 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -6682,14 +6454,8 @@ STAGE PLANS: #### A masked pattern was here #### location hdfs://### HDFS PATH ### name default.src5_n1 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src5_n1 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src5_n1 name: default.src5_n1 diff --git a/ql/src/test/results/clientpositive/llap/vectorization_0.q.out b/ql/src/test/results/clientpositive/llap/vectorization_0.q.out index 2c00a799d6..ad6b424ce7 100644 --- a/ql/src/test/results/clientpositive/llap/vectorization_0.q.out +++ b/ql/src/test/results/clientpositive/llap/vectorization_0.q.out @@ -1323,30 +1323,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 - columns.comments columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesorc - numFiles 1 - numRows 12288 - rawDataSize 2907994 - serialization.ddl struct alltypesorc { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 295616 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 @@ -1354,14 +1344,8 @@ STAGE PLANS: columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesorc - numFiles 1 - numRows 12288 - rawDataSize 2907994 - serialization.ddl struct alltypesorc { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 295616 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypesorc name: default.alltypesorc @@ -30158,30 +30142,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 - columns.comments columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesorc - numFiles 1 - numRows 12288 - rawDataSize 2907994 - serialization.ddl struct alltypesorc { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 295616 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 @@ -30189,14 +30163,8 @@ STAGE PLANS: columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesorc - numFiles 1 - numRows 12288 - rawDataSize 2907994 - serialization.ddl struct alltypesorc { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 295616 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypesorc name: default.alltypesorc @@ -30285,30 +30253,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 - columns.comments columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesorc - numFiles 1 - numRows 12288 - rawDataSize 2907994 - serialization.ddl struct alltypesorc { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 295616 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 @@ -30316,14 +30274,8 @@ STAGE PLANS: columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesorc - numFiles 1 - numRows 12288 - rawDataSize 2907994 - serialization.ddl struct alltypesorc { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 295616 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypesorc name: default.alltypesorc @@ -30412,30 +30364,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 - columns.comments columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesorc - numFiles 1 - numRows 12288 - rawDataSize 2907994 - serialization.ddl struct alltypesorc { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 295616 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 @@ -30443,14 +30385,8 @@ STAGE PLANS: columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesorc - numFiles 1 - numRows 12288 - rawDataSize 2907994 - serialization.ddl struct alltypesorc { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 295616 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypesorc name: default.alltypesorc @@ -30533,30 +30469,20 @@ STAGE PLANS: input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 - columns.comments columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesorc - numFiles 1 - numRows 12288 - rawDataSize 2907994 - serialization.ddl struct alltypesorc { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 295616 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde input format: org.apache.hadoop.hive.ql.io.orc.OrcInputFormat output format: org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"cbigint":"true","cboolean1":"true","cboolean2":"true","cdouble":"true","cfloat":"true","cint":"true","csmallint":"true","cstring1":"true","cstring2":"true","ctimestamp1":"true","ctimestamp2":"true","ctinyint":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns ctinyint,csmallint,cint,cbigint,cfloat,cdouble,cstring1,cstring2,ctimestamp1,ctimestamp2,cboolean1,cboolean2 @@ -30564,14 +30490,8 @@ STAGE PLANS: columns.types tinyint:smallint:int:bigint:float:double:string:string:timestamp:timestamp:boolean:boolean #### A masked pattern was here #### name default.alltypesorc - numFiles 1 - numRows 12288 - rawDataSize 2907994 - serialization.ddl struct alltypesorc { byte ctinyint, i16 csmallint, i32 cint, i64 cbigint, float cfloat, double cdouble, string cstring1, string cstring2, timestamp ctimestamp1, timestamp ctimestamp2, bool cboolean1, bool cboolean2} serialization.format 1 serialization.lib org.apache.hadoop.hive.ql.io.orc.OrcSerde - totalSize 295616 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.ql.io.orc.OrcSerde name: default.alltypesorc name: default.alltypesorc diff --git a/ql/src/test/results/clientpositive/regexp_extract.q.out b/ql/src/test/results/clientpositive/regexp_extract.q.out index 95f7c22bc9..fa0015f637 100644 --- a/ql/src/test/results/clientpositive/regexp_extract.q.out +++ b/ql/src/test/results/clientpositive/regexp_extract.q.out @@ -75,30 +75,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -106,14 +96,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -337,30 +321,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -368,14 +342,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/serde_user_properties.q.out b/ql/src/test/results/clientpositive/serde_user_properties.q.out index ac2b2ee6c9..b513b2b3c4 100644 --- a/ql/src/test/results/clientpositive/serde_user_properties.q.out +++ b/ql/src/test/results/clientpositive/serde_user_properties.q.out @@ -116,30 +116,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -147,14 +137,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -259,31 +243,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### user.defined.key some.value serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -291,14 +265,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### user.defined.key some.value serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src @@ -404,31 +372,21 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### user.defined.key some.value serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -436,14 +394,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### user.defined.key some.value serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src diff --git a/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out b/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out index 1142daba9c..8be3f9a673 100644 --- a/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out +++ b/ql/src/test/results/clientpositive/sort_merge_join_desc_5.q.out @@ -118,25 +118,17 @@ STAGE PLANS: partition values: part 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 1 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n7 - numFiles 1 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_1_n7 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -154,10 +146,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n7 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n7 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n7 name: default.srcbucket_mapjoin_part_1_n7 diff --git a/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out b/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out index 17f3b0b360..c13071811f 100644 --- a/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out +++ b/ql/src/test/results/clientpositive/sort_merge_join_desc_6.q.out @@ -85,25 +85,17 @@ STAGE PLANS: partition values: part 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n8 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_2_n8 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -120,10 +112,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n8 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n8 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n8 name: default.srcbucket_mapjoin_part_2_n8 @@ -201,25 +191,17 @@ STAGE PLANS: partition values: part 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n3 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_1_n3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -237,10 +219,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n3 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n3 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n3 name: default.srcbucket_mapjoin_part_1_n3 diff --git a/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out b/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out index 51bb46b399..c8a07db5d7 100644 --- a/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out +++ b/ql/src/test/results/clientpositive/sort_merge_join_desc_7.q.out @@ -125,25 +125,17 @@ STAGE PLANS: partition values: part 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key,value column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n2 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_2_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -160,10 +152,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n2 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n2 name: default.srcbucket_mapjoin_part_2_n2 @@ -173,25 +163,17 @@ STAGE PLANS: partition values: part 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key,value column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_2_n2 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_2_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -208,10 +190,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_2_n2 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_2_n2 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_2_n2 name: default.srcbucket_mapjoin_part_2_n2 @@ -281,25 +261,17 @@ STAGE PLANS: partition values: part 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key,value column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n0 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -316,10 +288,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n0 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n0 name: default.srcbucket_mapjoin_part_1_n0 @@ -331,25 +301,17 @@ STAGE PLANS: partition values: part 2 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count 2 bucket_field_name key,value column.name.delimiter , columns key,value - columns.comments columns.types int:string #### A masked pattern was here #### name default.srcbucket_mapjoin_part_1_n0 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 5312 - serialization.ddl struct srcbucket_mapjoin_part_1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat @@ -366,10 +328,8 @@ STAGE PLANS: name default.srcbucket_mapjoin_part_1_n0 partition_columns part partition_columns.types string - serialization.ddl struct srcbucket_mapjoin_part_1_n0 { i32 key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcbucket_mapjoin_part_1_n0 name: default.srcbucket_mapjoin_part_1_n0 diff --git a/ql/src/test/results/clientpositive/temp_table_partition_pruning.q.out b/ql/src/test/results/clientpositive/temp_table_partition_pruning.q.out index f6fdd61928..fe49987e3c 100644 --- a/ql/src/test/results/clientpositive/temp_table_partition_pruning.q.out +++ b/ql/src/test/results/clientpositive/temp_table_partition_pruning.q.out @@ -172,30 +172,20 @@ STAGE PLANS: partition values: dt 2001-01-01 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 column.name.delimiter , columns customer - columns.comments columns.types int #### A masked pattern was here #### name default.daysales_temp - numFiles 1 - numRows 1 partition_columns dt partition_columns.types string - rawDataSize 1 - serialization.ddl struct daysales_temp { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns customer @@ -203,15 +193,10 @@ STAGE PLANS: columns.types int #### A masked pattern was here #### name default.daysales_temp - numFiles 0 - numRows 0 partition_columns dt partition_columns.types string - rawDataSize 0 - serialization.ddl struct daysales_temp { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.daysales_temp name: default.daysales_temp @@ -223,30 +208,20 @@ STAGE PLANS: partition values: dt 2001-01-03 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 column.name.delimiter , columns customer - columns.comments columns.types int #### A masked pattern was here #### name default.daysales_temp - numFiles 1 - numRows 1 partition_columns dt partition_columns.types string - rawDataSize 1 - serialization.ddl struct daysales_temp { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns customer @@ -254,15 +229,10 @@ STAGE PLANS: columns.types int #### A masked pattern was here #### name default.daysales_temp - numFiles 0 - numRows 0 partition_columns dt partition_columns.types string - rawDataSize 0 - serialization.ddl struct daysales_temp { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.daysales_temp name: default.daysales_temp @@ -347,30 +317,20 @@ STAGE PLANS: partition values: dt 2001-01-01 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 column.name.delimiter , columns customer - columns.comments columns.types int #### A masked pattern was here #### name default.daysales_temp - numFiles 1 - numRows 1 partition_columns dt partition_columns.types string - rawDataSize 1 - serialization.ddl struct daysales_temp { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns customer @@ -378,15 +338,10 @@ STAGE PLANS: columns.types int #### A masked pattern was here #### name default.daysales_temp - numFiles 0 - numRows 0 partition_columns dt partition_columns.types string - rawDataSize 0 - serialization.ddl struct daysales_temp { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.daysales_temp name: default.daysales_temp @@ -398,30 +353,20 @@ STAGE PLANS: partition values: dt 2001-01-03 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 column.name.delimiter , columns customer - columns.comments columns.types int #### A masked pattern was here #### name default.daysales_temp - numFiles 1 - numRows 1 partition_columns dt partition_columns.types string - rawDataSize 1 - serialization.ddl struct daysales_temp { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns customer @@ -429,15 +374,10 @@ STAGE PLANS: columns.types int #### A masked pattern was here #### name default.daysales_temp - numFiles 0 - numRows 0 partition_columns dt partition_columns.types string - rawDataSize 0 - serialization.ddl struct daysales_temp { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.daysales_temp name: default.daysales_temp @@ -522,30 +462,20 @@ STAGE PLANS: partition values: dt 2001-01-01 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 column.name.delimiter , columns customer - columns.comments columns.types int #### A masked pattern was here #### name default.daysales_temp - numFiles 1 - numRows 1 partition_columns dt partition_columns.types string - rawDataSize 1 - serialization.ddl struct daysales_temp { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns customer @@ -553,15 +483,10 @@ STAGE PLANS: columns.types int #### A masked pattern was here #### name default.daysales_temp - numFiles 0 - numRows 0 partition_columns dt partition_columns.types string - rawDataSize 0 - serialization.ddl struct daysales_temp { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.daysales_temp name: default.daysales_temp @@ -573,30 +498,20 @@ STAGE PLANS: partition values: dt 2001-01-03 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 column.name.delimiter , columns customer - columns.comments columns.types int #### A masked pattern was here #### name default.daysales_temp - numFiles 1 - numRows 1 partition_columns dt partition_columns.types string - rawDataSize 1 - serialization.ddl struct daysales_temp { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 2 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"customer":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns customer @@ -604,15 +519,10 @@ STAGE PLANS: columns.types int #### A masked pattern was here #### name default.daysales_temp - numFiles 0 - numRows 0 partition_columns dt partition_columns.types string - rawDataSize 0 - serialization.ddl struct daysales_temp { i32 customer} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 0 serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.daysales_temp name: default.daysales_temp diff --git a/ql/src/test/results/clientpositive/timestamp.q.out b/ql/src/test/results/clientpositive/timestamp.q.out index 90a46f58f4..2994fb8e57 100644 --- a/ql/src/test/results/clientpositive/timestamp.q.out +++ b/ql/src/test/results/clientpositive/timestamp.q.out @@ -164,30 +164,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -195,14 +185,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/transform_ppr1.q.out b/ql/src/test/results/clientpositive/transform_ppr1.q.out index 25468bcd9c..120cbfe328 100644 --- a/ql/src/test/results/clientpositive/transform_ppr1.q.out +++ b/ql/src/test/results/clientpositive/transform_ppr1.q.out @@ -87,30 +87,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -120,10 +110,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -136,30 +124,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -169,10 +147,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -185,30 +161,20 @@ STAGE PLANS: ds 2008-04-09 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -218,10 +184,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -234,30 +198,20 @@ STAGE PLANS: ds 2008-04-09 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -267,10 +221,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/transform_ppr2.q.out b/ql/src/test/results/clientpositive/transform_ppr2.q.out index 8aeb688513..a95308b327 100644 --- a/ql/src/test/results/clientpositive/transform_ppr2.q.out +++ b/ql/src/test/results/clientpositive/transform_ppr2.q.out @@ -86,30 +86,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -119,10 +109,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -135,30 +123,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -168,10 +146,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/ql/src/test/results/clientpositive/truncate_column_list_bucket.q.out b/ql/src/test/results/clientpositive/truncate_column_list_bucket.q.out index c8e40bd447..d30bb93dbd 100644 --- a/ql/src/test/results/clientpositive/truncate_column_list_bucket.q.out +++ b/ql/src/test/results/clientpositive/truncate_column_list_bucket.q.out @@ -119,29 +119,20 @@ STAGE PLANS: partition values: part 1 properties: - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.test_tab_n3 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 4812 - serialization.ddl struct test_tab_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 1761 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -151,10 +142,8 @@ STAGE PLANS: name default.test_tab_n3 partition_columns part partition_columns.types string - serialization.ddl struct test_tab_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.test_tab_n3 name: default.test_tab_n3 @@ -248,29 +237,20 @@ STAGE PLANS: partition values: part 1 properties: - bucket_count -1 column.name.delimiter , columns key,value - columns.comments columns.types string:string #### A masked pattern was here #### name default.test_tab_n3 - numFiles 2 - numRows 500 partition_columns part partition_columns.types string - rawDataSize 4812 - serialization.ddl struct test_tab_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe - totalSize 1761 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe input format: org.apache.hadoop.hive.ql.io.RCFileInputFormat output format: org.apache.hadoop.hive.ql.io.RCFileOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -280,10 +260,8 @@ STAGE PLANS: name default.test_tab_n3 partition_columns part partition_columns.types string - serialization.ddl struct test_tab_n3 { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.columnar.ColumnarSerDe name: default.test_tab_n3 name: default.test_tab_n3 diff --git a/ql/src/test/results/clientpositive/udf_explode.q.out b/ql/src/test/results/clientpositive/udf_explode.q.out index 0143f3160b..a679788f46 100644 --- a/ql/src/test/results/clientpositive/udf_explode.q.out +++ b/ql/src/test/results/clientpositive/udf_explode.q.out @@ -95,30 +95,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -126,14 +116,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -296,30 +280,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -327,14 +301,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/udtf_explode.q.out b/ql/src/test/results/clientpositive/udtf_explode.q.out index 1b941b87bb..764673bf15 100644 --- a/ql/src/test/results/clientpositive/udtf_explode.q.out +++ b/ql/src/test/results/clientpositive/udtf_explode.q.out @@ -101,30 +101,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -132,14 +122,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src @@ -387,30 +371,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -418,14 +392,8 @@ STAGE PLANS: columns.types string:string #### A masked pattern was here #### name default.src - numFiles 1 - numRows 500 - rawDataSize 5312 - serialization.ddl struct src { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src name: default.src diff --git a/ql/src/test/results/clientpositive/union22.q.out b/ql/src/test/results/clientpositive/union22.q.out index de36e44dfb..d262942b9c 100644 --- a/ql/src/test/results/clientpositive/union22.q.out +++ b/ql/src/test/results/clientpositive/union22.q.out @@ -118,30 +118,20 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"k0":"true","k1":"true","k2":"true","k3":"true","k4":"true","k5":"true"}} - bucket_count -1 column.name.delimiter , columns k0,k1,k2,k3,k4,k5 - columns.comments columns.types string:string:string:string:string:string #### A masked pattern was here #### name default.dst_union22_delta - numFiles 1 - numRows 500 partition_columns ds partition_columns.types string - rawDataSize 16936 - serialization.ddl struct dst_union22_delta { string k0, string k1, string k2, string k3, string k4, string k5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17436 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k0,k1,k2,k3,k4,k5 @@ -151,10 +141,8 @@ STAGE PLANS: name default.dst_union22_delta partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22_delta { string k0, string k1, string k2, string k3, string k4, string k5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22_delta name: default.dst_union22_delta @@ -251,30 +239,20 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"k1":"true","k2":"true","k3":"true","k4":"true"}} - bucket_count -1 column.name.delimiter , columns k1,k2,k3,k4 - columns.comments columns.types string:string:string:string #### A masked pattern was here #### name default.dst_union22 - numFiles 1 - numRows 500 partition_columns ds partition_columns.types string - rawDataSize 11124 - serialization.ddl struct dst_union22 { string k1, string k2, string k3, string k4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k1,k2,k3,k4 @@ -284,10 +262,8 @@ STAGE PLANS: name default.dst_union22 partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22 { string k1, string k2, string k3, string k4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22 name: default.dst_union22 @@ -299,30 +275,20 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"k0":"true","k1":"true","k2":"true","k3":"true","k4":"true","k5":"true"}} - bucket_count -1 column.name.delimiter , columns k0,k1,k2,k3,k4,k5 - columns.comments columns.types string:string:string:string:string:string #### A masked pattern was here #### name default.dst_union22_delta - numFiles 1 - numRows 500 partition_columns ds partition_columns.types string - rawDataSize 16936 - serialization.ddl struct dst_union22_delta { string k0, string k1, string k2, string k3, string k4, string k5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17436 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k0,k1,k2,k3,k4,k5 @@ -332,10 +298,8 @@ STAGE PLANS: name default.dst_union22_delta partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22_delta { string k0, string k1, string k2, string k3, string k4, string k5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22_delta name: default.dst_union22_delta @@ -373,7 +337,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k1,k2,k3,k4 @@ -383,10 +346,8 @@ STAGE PLANS: name default.dst_union22 partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22 { string k1, string k2, string k3, string k4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22 TotalFiles: 1 @@ -431,7 +392,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k1,k2,k3,k4 @@ -441,10 +401,8 @@ STAGE PLANS: name default.dst_union22 partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22 { string k1, string k2, string k3, string k4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22 TotalFiles: 1 @@ -505,30 +463,20 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"k0":"true","k1":"true","k2":"true","k3":"true","k4":"true","k5":"true"}} - bucket_count -1 column.name.delimiter , columns k0,k1,k2,k3,k4,k5 - columns.comments columns.types string:string:string:string:string:string #### A masked pattern was here #### name default.dst_union22_delta - numFiles 1 - numRows 500 partition_columns ds partition_columns.types string - rawDataSize 16936 - serialization.ddl struct dst_union22_delta { string k0, string k1, string k2, string k3, string k4, string k5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17436 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k0,k1,k2,k3,k4,k5 @@ -538,10 +486,8 @@ STAGE PLANS: name default.dst_union22_delta partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22_delta { string k0, string k1, string k2, string k3, string k4, string k5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22_delta name: default.dst_union22_delta @@ -596,7 +542,6 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k1,k2,k3,k4 @@ -606,10 +551,8 @@ STAGE PLANS: name default.dst_union22 partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22 { string k1, string k2, string k3, string k4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22 @@ -685,30 +628,20 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"k1":"true","k2":"true","k3":"true","k4":"true"}} - bucket_count -1 column.name.delimiter , columns k1,k2,k3,k4 - columns.comments columns.types string:string:string:string #### A masked pattern was here #### name default.dst_union22 - numFiles 1 - numRows 500 partition_columns ds partition_columns.types string - rawDataSize 11124 - serialization.ddl struct dst_union22 { string k1, string k2, string k3, string k4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 11624 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k1,k2,k3,k4 @@ -718,10 +651,8 @@ STAGE PLANS: name default.dst_union22 partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22 { string k1, string k2, string k3, string k4} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22 name: default.dst_union22 @@ -733,30 +664,20 @@ STAGE PLANS: partition values: ds 1 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"k0":"true","k1":"true","k2":"true","k3":"true","k4":"true","k5":"true"}} - bucket_count -1 column.name.delimiter , columns k0,k1,k2,k3,k4,k5 - columns.comments columns.types string:string:string:string:string:string #### A masked pattern was here #### name default.dst_union22_delta - numFiles 1 - numRows 500 partition_columns ds partition_columns.types string - rawDataSize 16936 - serialization.ddl struct dst_union22_delta { string k0, string k1, string k2, string k3, string k4, string k5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 17436 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns k0,k1,k2,k3,k4,k5 @@ -766,10 +687,8 @@ STAGE PLANS: name default.dst_union22_delta partition_columns ds partition_columns.types string - serialization.ddl struct dst_union22_delta { string k0, string k1, string k2, string k3, string k4, string k5} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.dst_union22_delta name: default.dst_union22_delta diff --git a/ql/src/test/results/clientpositive/union24.q.out b/ql/src/test/results/clientpositive/union24.q.out index 32a86e7f02..be1c232844 100644 --- a/ql/src/test/results/clientpositive/union24.q.out +++ b/ql/src/test/results/clientpositive/union24.q.out @@ -140,30 +140,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### name default.src5_n3 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src5_n3 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -171,14 +161,8 @@ STAGE PLANS: columns.types string:bigint #### A masked pattern was here #### name default.src5_n3 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src5_n3 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src5_n3 name: default.src5_n3 @@ -391,30 +375,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### name default.src2_n6 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src2_n6 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -422,14 +396,8 @@ STAGE PLANS: columns.types string:bigint #### A masked pattern was here #### name default.src2_n6 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src2_n6 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src2_n6 name: default.src2_n6 @@ -439,30 +407,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### name default.src3_n2 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src3_n2 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -470,14 +428,8 @@ STAGE PLANS: columns.types string:bigint #### A masked pattern was here #### name default.src3_n2 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src3_n2 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src3_n2 name: default.src3_n2 @@ -487,30 +439,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### name default.src4_n0 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src4_n0 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -518,14 +460,8 @@ STAGE PLANS: columns.types string:bigint #### A masked pattern was here #### name default.src4_n0 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src4_n0 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src4_n0 name: default.src4_n0 @@ -703,30 +639,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### name default.src4_n0 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src4_n0 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -734,14 +660,8 @@ STAGE PLANS: columns.types string:bigint #### A masked pattern was here #### name default.src4_n0 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src4_n0 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src4_n0 name: default.src4_n0 @@ -751,30 +671,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### name default.src5_n3 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src5_n3 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -782,14 +692,8 @@ STAGE PLANS: columns.types string:bigint #### A masked pattern was here #### name default.src5_n3 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src5_n3 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src5_n3 name: default.src5_n3 @@ -970,30 +874,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### name default.src2_n6 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src2_n6 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -1001,14 +895,8 @@ STAGE PLANS: columns.types string:bigint #### A masked pattern was here #### name default.src2_n6 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src2_n6 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src2_n6 name: default.src2_n6 @@ -1018,30 +906,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### name default.src3_n2 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src3_n2 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -1049,14 +927,8 @@ STAGE PLANS: columns.types string:bigint #### A masked pattern was here #### name default.src3_n2 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src3_n2 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src3_n2 name: default.src3_n2 @@ -1224,30 +1096,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### name default.src4_n0 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src4_n0 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -1255,14 +1117,8 @@ STAGE PLANS: columns.types string:bigint #### A masked pattern was here #### name default.src4_n0 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src4_n0 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src4_n0 name: default.src4_n0 @@ -1272,30 +1128,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### name default.src5_n3 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src5_n3 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -1303,14 +1149,8 @@ STAGE PLANS: columns.types string:bigint #### A masked pattern was here #### name default.src5_n3 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src5_n3 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src5_n3 name: default.src5_n3 @@ -1566,30 +1406,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### name default.src2_n6 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src2_n6 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -1597,14 +1427,8 @@ STAGE PLANS: columns.types string:bigint #### A masked pattern was here #### name default.src2_n6 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src2_n6 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src2_n6 name: default.src2_n6 @@ -1614,30 +1438,20 @@ STAGE PLANS: input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count - columns.comments columns.types string:bigint #### A masked pattern was here #### name default.src3_n2 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src3_n2 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"count":"true","key":"true"}} - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,count @@ -1645,14 +1459,8 @@ STAGE PLANS: columns.types string:bigint #### A masked pattern was here #### name default.src3_n2 - numFiles 1 - numRows 309 - rawDataSize 1482 - serialization.ddl struct src3_n2 { string key, i64 count} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 1791 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.src3_n2 name: default.src3_n2 diff --git a/ql/src/test/results/clientpositive/union_ppr.q.out b/ql/src/test/results/clientpositive/union_ppr.q.out index b841994373..168b75f4af 100644 --- a/ql/src/test/results/clientpositive/union_ppr.q.out +++ b/ql/src/test/results/clientpositive/union_ppr.q.out @@ -100,30 +100,20 @@ STAGE PLANS: ds 2008-04-08 hr 11 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -133,10 +123,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart @@ -149,30 +137,20 @@ STAGE PLANS: ds 2008-04-08 hr 12 properties: - COLUMN_STATS_ACCURATE {"BASIC_STATS":"true","COLUMN_STATS":{"key":"true","value":"true"}} - bucket_count -1 column.name.delimiter , columns key,value - columns.comments 'default','default' columns.types string:string #### A masked pattern was here #### name default.srcpart - numFiles 1 - numRows 500 partition_columns ds/hr partition_columns.types string:string - rawDataSize 5312 - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe - totalSize 5812 -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe input format: org.apache.hadoop.mapred.TextInputFormat output format: org.apache.hadoop.hive.ql.io.HiveIgnoreKeyTextOutputFormat properties: - bucket_count -1 bucketing_version 2 column.name.delimiter , columns key,value @@ -182,10 +160,8 @@ STAGE PLANS: name default.srcpart partition_columns ds/hr partition_columns.types string:string - serialization.ddl struct srcpart { string key, string value} serialization.format 1 serialization.lib org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe -#### A masked pattern was here #### serde: org.apache.hadoop.hive.serde2.lazy.LazySimpleSerDe name: default.srcpart name: default.srcpart diff --git a/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type/DynamicSerDe.java b/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type/DynamicSerDe.java index 2b832ac436..a6ed8b8fe1 100644 --- a/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type/DynamicSerDe.java +++ b/serde/src/java/org/apache/hadoop/hive/serde2/dynamic_type/DynamicSerDe.java @@ -20,6 +20,7 @@ import java.io.ByteArrayInputStream; import java.util.ArrayList; +import java.util.HashMap; import java.util.List; import java.util.Properties; @@ -77,6 +78,83 @@ TIOStreamTransport tios; + private static HashMap typeToThriftTypeMap; + static { + typeToThriftTypeMap = new HashMap<>(); + typeToThriftTypeMap.put(serdeConstants.BOOLEAN_TYPE_NAME, "bool"); + typeToThriftTypeMap.put(serdeConstants.TINYINT_TYPE_NAME, "byte"); + typeToThriftTypeMap.put(serdeConstants.SMALLINT_TYPE_NAME, "i16"); + typeToThriftTypeMap.put(serdeConstants.INT_TYPE_NAME, "i32"); + typeToThriftTypeMap.put(serdeConstants.BIGINT_TYPE_NAME, "i64"); + typeToThriftTypeMap.put(serdeConstants.DOUBLE_TYPE_NAME, "double"); + typeToThriftTypeMap.put(serdeConstants.FLOAT_TYPE_NAME, "float"); + typeToThriftTypeMap.put(serdeConstants.LIST_TYPE_NAME, "list"); + typeToThriftTypeMap.put(serdeConstants.MAP_TYPE_NAME, "map"); + typeToThriftTypeMap.put(serdeConstants.STRING_TYPE_NAME, "string"); + typeToThriftTypeMap.put(serdeConstants.BINARY_TYPE_NAME, "binary"); + // These 4 types are not supported yet. + // We should define a complex type date in thrift that contains a single int + // member, and DynamicSerDe + // should convert it to date type at runtime. + typeToThriftTypeMap.put(serdeConstants.DATE_TYPE_NAME, "date"); + typeToThriftTypeMap.put(serdeConstants.DATETIME_TYPE_NAME, "datetime"); + typeToThriftTypeMap.put(serdeConstants.TIMESTAMP_TYPE_NAME, "timestamp"); + typeToThriftTypeMap.put(serdeConstants.DECIMAL_TYPE_NAME, "decimal"); + } + + /** + * Convert type to ThriftType. We do that by tokenizing the type and convert + * each token. + */ + private String typeToThriftType(String type) { + StringBuilder thriftType = new StringBuilder(); + int last = 0; + boolean lastAlphaDigit = Character.isLetterOrDigit(type.charAt(last)); + for (int i = 1; i <= type.length(); i++) { + if (i == type.length() + || Character.isLetterOrDigit(type.charAt(i)) != lastAlphaDigit) { + String token = type.substring(last, i); + last = i; + String thriftToken = typeToThriftTypeMap.get(token); + thriftType.append(thriftToken == null ? token : thriftToken); + lastAlphaDigit = !lastAlphaDigit; + } + } + return thriftType.toString(); + } + + /** + * Generate Thrift DDL from column props. + */ + private String getDDLFrom(String structName, + String cols, String types) throws SerDeException{ + String[] colNames = cols.split(","); + String[] colTypes = types.split(":"); + + if (colNames.length != colTypes.length) { + throw new SerDeException("colnames : " + cols + " types: "+types); + } + StringBuilder ddl = new StringBuilder(); + ddl.append("struct "); + ddl.append(structName); + ddl.append(" { "); + boolean first = true; + for (int i=0; i < colNames.length; i++) { + if (first) { + first = false; + } else { + ddl.append(", "); + } + ddl.append(typeToThriftType(colTypes[i])); + ddl.append(' '); + ddl.append(colNames[i]); + } + ddl.append("}"); + + LOG.trace("DDL: {}", ddl); + return ddl.toString(); + } + @Override public void initialize(Configuration job, Properties tbl) throws SerDeException { try { @@ -93,6 +171,11 @@ public void initialize(Configuration job, Properties tbl) throws SerDeException } else { type_name = tableName; } + + if (null == ddl) { + ddl = getDDLFrom(type_name, tbl.getProperty(serdeConstants.LIST_COLUMNS), tbl.getProperty(serdeConstants.LIST_COLUMN_TYPES)); + + } String protoName = tbl.getProperty(serdeConstants.SERIALIZATION_FORMAT); if (protoName == null) { diff --git a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/MetaStoreUtils.java b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/MetaStoreUtils.java index 18f689ebf4..a70a8fa403 100644 --- a/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/MetaStoreUtils.java +++ b/standalone-metastore/metastore-common/src/main/java/org/apache/hadoop/hive/metastore/utils/MetaStoreUtils.java @@ -43,6 +43,7 @@ import org.apache.hadoop.conf.Configuration; import org.apache.hadoop.fs.CommonConfigurationKeysPublic; import org.apache.hadoop.fs.Path; +import org.apache.hadoop.hive.common.StatsSetupConst; import org.apache.hadoop.hive.metastore.ColumnType; import org.apache.hadoop.hive.metastore.TableType; import org.apache.hadoop.hive.metastore.Warehouse; @@ -661,9 +662,12 @@ public static Properties getSchemaWithoutCols(StorageDescriptor sd, org.apache.hadoop.hive.metastore.api.hive_metastoreConstants.META_TABLE_LOCATION, sd.getLocation()); } - schema.setProperty( - org.apache.hadoop.hive.metastore.api.hive_metastoreConstants.BUCKET_COUNT, Integer - .toString(sd.getNumBuckets())); + int bucket_cnt = sd.getNumBuckets(); + if (bucket_cnt > 0) { + schema.setProperty(org.apache.hadoop.hive.metastore.api.hive_metastoreConstants.BUCKET_COUNT, + Integer.toString(bucket_cnt)); + } + if (sd.getBucketCols() != null && sd.getBucketCols().size() > 0) { schema.setProperty( org.apache.hadoop.hive.metastore.api.hive_metastoreConstants.BUCKET_FIELD_NAME, @@ -679,10 +683,6 @@ public static Properties getSchemaWithoutCols(StorageDescriptor sd, } } - if (sd.getCols() != null) { - schema.setProperty(ColumnType.SERIALIZATION_DDL, getDDLFromFieldSchema(tableName, sd.getCols())); - } - String partString = StringUtils.EMPTY; String partStringSep = StringUtils.EMPTY; String partTypesString = StringUtils.EMPTY; @@ -711,7 +711,13 @@ public static Properties getSchemaWithoutCols(StorageDescriptor sd, if (parameters != null) { for (Map.Entry e : parameters.entrySet()) { // add non-null parameters to the schema - if ( e.getValue() != null) { + String key = e.getKey(); + if (!StatsSetupConst.COLUMN_STATS_ACCURATE.equals(key) && + !hive_metastoreConstants.DDL_TIME.equals(key) && + !StatsSetupConst.TOTAL_SIZE.equals(key) && + !StatsSetupConst.RAW_DATA_SIZE.equals(key) && + !StatsSetupConst.NUM_FILES.equals(key) && + !StatsSetupConst.ROW_COUNT.equals(key) && e.getValue() != null) { schema.setProperty(e.getKey(), e.getValue()); } } -- 2.17.2 (Apple Git-113)